Dec 06 06:53:05 crc systemd[1]: Starting Kubernetes Kubelet... Dec 06 06:53:05 crc restorecon[4758]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:05 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:06 crc restorecon[4758]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 06:53:06 crc restorecon[4758]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 06 06:53:06 crc kubenswrapper[4945]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 06:53:06 crc kubenswrapper[4945]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 06 06:53:06 crc kubenswrapper[4945]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 06:53:06 crc kubenswrapper[4945]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 06:53:06 crc kubenswrapper[4945]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 06 06:53:06 crc kubenswrapper[4945]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.766016 4945 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.773890 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.773924 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.773937 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.773950 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.773961 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.773972 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.773987 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774002 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774014 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774025 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774035 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774047 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774058 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774070 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774080 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774090 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774101 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774114 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774126 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774140 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774154 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774166 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774176 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774187 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774198 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774208 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774219 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774229 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774239 4945 feature_gate.go:330] unrecognized feature gate: Example Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774251 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774261 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774327 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774343 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774354 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774364 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774379 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774391 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774403 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774419 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774429 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774440 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774452 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774462 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774472 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774481 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774491 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774502 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774511 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774521 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774530 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774540 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774550 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774560 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774572 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774582 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774593 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774603 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774613 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774623 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774632 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774643 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774654 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774664 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774677 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774688 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774698 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774708 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774737 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774749 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774758 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.774768 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775335 4945 flags.go:64] FLAG: --address="0.0.0.0" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775368 4945 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775402 4945 flags.go:64] FLAG: --anonymous-auth="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775417 4945 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775432 4945 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775446 4945 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775467 4945 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775481 4945 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775494 4945 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775507 4945 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775520 4945 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775533 4945 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775544 4945 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775561 4945 flags.go:64] FLAG: --cgroup-root="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775573 4945 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775585 4945 flags.go:64] FLAG: --client-ca-file="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775597 4945 flags.go:64] FLAG: --cloud-config="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775608 4945 flags.go:64] FLAG: --cloud-provider="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775620 4945 flags.go:64] FLAG: --cluster-dns="[]" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775661 4945 flags.go:64] FLAG: --cluster-domain="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775672 4945 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775684 4945 flags.go:64] FLAG: --config-dir="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775695 4945 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775708 4945 flags.go:64] FLAG: --container-log-max-files="5" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775722 4945 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775736 4945 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775749 4945 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775761 4945 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775772 4945 flags.go:64] FLAG: --contention-profiling="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775783 4945 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775795 4945 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775807 4945 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775838 4945 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775853 4945 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775866 4945 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775878 4945 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775890 4945 flags.go:64] FLAG: --enable-load-reader="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775903 4945 flags.go:64] FLAG: --enable-server="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775915 4945 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775940 4945 flags.go:64] FLAG: --event-burst="100" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775953 4945 flags.go:64] FLAG: --event-qps="50" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775966 4945 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775978 4945 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.775989 4945 flags.go:64] FLAG: --eviction-hard="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776004 4945 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776018 4945 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776029 4945 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776042 4945 flags.go:64] FLAG: --eviction-soft="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776053 4945 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776064 4945 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776075 4945 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776086 4945 flags.go:64] FLAG: --experimental-mounter-path="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776098 4945 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776109 4945 flags.go:64] FLAG: --fail-swap-on="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776120 4945 flags.go:64] FLAG: --feature-gates="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776134 4945 flags.go:64] FLAG: --file-check-frequency="20s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776145 4945 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776158 4945 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776171 4945 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776181 4945 flags.go:64] FLAG: --healthz-port="10248" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776192 4945 flags.go:64] FLAG: --help="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776202 4945 flags.go:64] FLAG: --hostname-override="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776213 4945 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776225 4945 flags.go:64] FLAG: --http-check-frequency="20s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776235 4945 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776246 4945 flags.go:64] FLAG: --image-credential-provider-config="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776256 4945 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776267 4945 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776332 4945 flags.go:64] FLAG: --image-service-endpoint="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776347 4945 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776358 4945 flags.go:64] FLAG: --kube-api-burst="100" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776370 4945 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776381 4945 flags.go:64] FLAG: --kube-api-qps="50" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776393 4945 flags.go:64] FLAG: --kube-reserved="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776405 4945 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776415 4945 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776428 4945 flags.go:64] FLAG: --kubelet-cgroups="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776440 4945 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776451 4945 flags.go:64] FLAG: --lock-file="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776462 4945 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776474 4945 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776486 4945 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776504 4945 flags.go:64] FLAG: --log-json-split-stream="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776555 4945 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776569 4945 flags.go:64] FLAG: --log-text-split-stream="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776580 4945 flags.go:64] FLAG: --logging-format="text" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776591 4945 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776603 4945 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776615 4945 flags.go:64] FLAG: --manifest-url="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776628 4945 flags.go:64] FLAG: --manifest-url-header="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776643 4945 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776654 4945 flags.go:64] FLAG: --max-open-files="1000000" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776668 4945 flags.go:64] FLAG: --max-pods="110" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776680 4945 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776692 4945 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776703 4945 flags.go:64] FLAG: --memory-manager-policy="None" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776714 4945 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776725 4945 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776757 4945 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776769 4945 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776794 4945 flags.go:64] FLAG: --node-status-max-images="50" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776805 4945 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776817 4945 flags.go:64] FLAG: --oom-score-adj="-999" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776829 4945 flags.go:64] FLAG: --pod-cidr="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776863 4945 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776880 4945 flags.go:64] FLAG: --pod-manifest-path="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776892 4945 flags.go:64] FLAG: --pod-max-pids="-1" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776904 4945 flags.go:64] FLAG: --pods-per-core="0" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776915 4945 flags.go:64] FLAG: --port="10250" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776927 4945 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776939 4945 flags.go:64] FLAG: --provider-id="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776950 4945 flags.go:64] FLAG: --qos-reserved="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776961 4945 flags.go:64] FLAG: --read-only-port="10255" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776972 4945 flags.go:64] FLAG: --register-node="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776983 4945 flags.go:64] FLAG: --register-schedulable="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.776995 4945 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777014 4945 flags.go:64] FLAG: --registry-burst="10" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777026 4945 flags.go:64] FLAG: --registry-qps="5" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777037 4945 flags.go:64] FLAG: --reserved-cpus="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777048 4945 flags.go:64] FLAG: --reserved-memory="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777063 4945 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777074 4945 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777087 4945 flags.go:64] FLAG: --rotate-certificates="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777099 4945 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777110 4945 flags.go:64] FLAG: --runonce="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777123 4945 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777134 4945 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777147 4945 flags.go:64] FLAG: --seccomp-default="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777159 4945 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777170 4945 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777182 4945 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777193 4945 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777205 4945 flags.go:64] FLAG: --storage-driver-password="root" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777217 4945 flags.go:64] FLAG: --storage-driver-secure="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777228 4945 flags.go:64] FLAG: --storage-driver-table="stats" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777240 4945 flags.go:64] FLAG: --storage-driver-user="root" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777251 4945 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777262 4945 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777276 4945 flags.go:64] FLAG: --system-cgroups="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777325 4945 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777363 4945 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777376 4945 flags.go:64] FLAG: --tls-cert-file="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777388 4945 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777408 4945 flags.go:64] FLAG: --tls-min-version="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777420 4945 flags.go:64] FLAG: --tls-private-key-file="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777431 4945 flags.go:64] FLAG: --topology-manager-policy="none" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777443 4945 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777455 4945 flags.go:64] FLAG: --topology-manager-scope="container" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777467 4945 flags.go:64] FLAG: --v="2" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777481 4945 flags.go:64] FLAG: --version="false" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777495 4945 flags.go:64] FLAG: --vmodule="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777509 4945 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.777520 4945 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.777936 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.777968 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.777981 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.777996 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778009 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778023 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778035 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778046 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778057 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778067 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778078 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778088 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778098 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778109 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778120 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778130 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778140 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778151 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778161 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778172 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778181 4945 feature_gate.go:330] unrecognized feature gate: Example Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778191 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778201 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778228 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778239 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778250 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778260 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778269 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778327 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778340 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778350 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778364 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778378 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778397 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778409 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778420 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778433 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778444 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778454 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778464 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778475 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778487 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778497 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778507 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778517 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778528 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778538 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778548 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778559 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778569 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778579 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778590 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778600 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778610 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778620 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778630 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778640 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778650 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778660 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778712 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778724 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778736 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778746 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778757 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778776 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778792 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778802 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778825 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778835 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778845 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.778855 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.778873 4945 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.794972 4945 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.795037 4945 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795175 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795191 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795202 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795212 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795221 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795231 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795241 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795252 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795260 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795272 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795312 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795322 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795331 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795340 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795350 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795360 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795368 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795376 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795385 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795394 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795402 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795410 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795419 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795428 4945 feature_gate.go:330] unrecognized feature gate: Example Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795436 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795445 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795456 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795468 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795478 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795488 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795500 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795518 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795527 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795537 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795547 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795555 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795565 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795574 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795583 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795593 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795602 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795611 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795620 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795630 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795642 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795654 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795664 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795674 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795683 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795693 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795702 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795710 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795719 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795728 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795737 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795746 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795754 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795762 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795770 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795779 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795788 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795796 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795805 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795813 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795822 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795830 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795838 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795846 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795855 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795863 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.795872 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.795888 4945 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796144 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796158 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796170 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796182 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796191 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796201 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796209 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796220 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796230 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796239 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796248 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796256 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796265 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796273 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796307 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796317 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796326 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796334 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796342 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796351 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796359 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796368 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796376 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796385 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796394 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796402 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796410 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796419 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796427 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796435 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796444 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796452 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796461 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796469 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796478 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796487 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796495 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796503 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796512 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796520 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796549 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796558 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796568 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796579 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796588 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796596 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796605 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796613 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796621 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796630 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796639 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796647 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796656 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796664 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796672 4945 feature_gate.go:330] unrecognized feature gate: Example Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796681 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796689 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796700 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796710 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796718 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796727 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796736 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796745 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796754 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796762 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796771 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796779 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796790 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796801 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796812 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.796822 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.796836 4945 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.797085 4945 server.go:940] "Client rotation is on, will bootstrap in background" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.801860 4945 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.801948 4945 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.802406 4945 server.go:997] "Starting client certificate rotation" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.802426 4945 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.802634 4945 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-16 07:57:30.220794746 +0000 UTC Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.802775 4945 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 241h4m23.418023958s for next certificate rotation Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.809531 4945 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.811756 4945 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.821499 4945 log.go:25] "Validated CRI v1 runtime API" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.840390 4945 log.go:25] "Validated CRI v1 image API" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.842168 4945 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.845817 4945 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-06-06-47-50-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.845870 4945 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.872905 4945 manager.go:217] Machine: {Timestamp:2025-12-06 06:53:06.870674448 +0000 UTC m=+0.325535532 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:4af87963-9c01-463b-9f2d-189a21242b1b BootID:15bb9f62-3356-4dc4-bade-82cf13a5e0d3 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e0:f7:7d Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e0:f7:7d Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:13:f4:b0 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:bb:42:11 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:c9:8a:8c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:2f:b7:1c Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:c8:51:91 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:da:10:2c:57:8f:d0 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:e6:4b:49:d9:62:71 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.873405 4945 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.873718 4945 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.875052 4945 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.875499 4945 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.875567 4945 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.875937 4945 topology_manager.go:138] "Creating topology manager with none policy" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.875961 4945 container_manager_linux.go:303] "Creating device plugin manager" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.876243 4945 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.876326 4945 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.876796 4945 state_mem.go:36] "Initialized new in-memory state store" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.876959 4945 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.878921 4945 kubelet.go:418] "Attempting to sync node with API server" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.878950 4945 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.878977 4945 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.878990 4945 kubelet.go:324] "Adding apiserver pod source" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.879003 4945 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.881023 4945 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.881472 4945 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.882621 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.882675 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.882753 4945 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 06 06:53:06 crc kubenswrapper[4945]: E1206 06:53:06.882758 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.220:6443: connect: connection refused" logger="UnhandledError" Dec 06 06:53:06 crc kubenswrapper[4945]: E1206 06:53:06.882792 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.220:6443: connect: connection refused" logger="UnhandledError" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883383 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883415 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883424 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883433 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883446 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883462 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883473 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883488 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883498 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883509 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883521 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.883528 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.884000 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.884576 4945 server.go:1280] "Started kubelet" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.884997 4945 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.885000 4945 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.885095 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.886230 4945 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 06 06:53:06 crc systemd[1]: Started Kubernetes Kubelet. Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.886896 4945 server.go:460] "Adding debug handlers to kubelet server" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.887389 4945 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.887420 4945 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.887578 4945 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.887590 4945 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.887557 4945 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 15:26:36.651699682 +0000 UTC Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.887662 4945 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 680h33m29.764042551s for next certificate rotation Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.887666 4945 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 06 06:53:06 crc kubenswrapper[4945]: E1206 06:53:06.887981 4945 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 06:53:06 crc kubenswrapper[4945]: E1206 06:53:06.889001 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="200ms" Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.890829 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:06 crc kubenswrapper[4945]: E1206 06:53:06.888070 4945 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.220:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e8dc651668021 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 06:53:06.884526113 +0000 UTC m=+0.339387177,LastTimestamp:2025-12-06 06:53:06.884526113 +0000 UTC m=+0.339387177,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 06:53:06 crc kubenswrapper[4945]: E1206 06:53:06.891006 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.220:6443: connect: connection refused" logger="UnhandledError" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.893031 4945 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.893064 4945 factory.go:55] Registering systemd factory Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.893077 4945 factory.go:221] Registration of the systemd container factory successfully Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.893973 4945 factory.go:153] Registering CRI-O factory Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.894016 4945 factory.go:221] Registration of the crio container factory successfully Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.894050 4945 factory.go:103] Registering Raw factory Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.894070 4945 manager.go:1196] Started watching for new ooms in manager Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.898635 4945 manager.go:319] Starting recovery of all containers Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900517 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900576 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900607 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900621 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900632 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900643 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900654 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900665 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900679 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900690 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900704 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900758 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900772 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900801 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900814 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900833 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900844 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900856 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900866 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900877 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900888 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900898 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900934 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900946 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900956 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900969 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900982 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.900993 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901004 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901018 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901030 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901040 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901051 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901083 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901096 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901106 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901116 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901127 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901136 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901147 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901158 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901172 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901183 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901193 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901205 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901217 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901230 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901241 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901251 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901261 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901287 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901299 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901316 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901328 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901339 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901351 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901363 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901375 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901387 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901422 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901433 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901444 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901457 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901468 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901480 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901492 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901503 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901515 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901527 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901539 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901550 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901561 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901572 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901585 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901598 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901608 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901620 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901629 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901639 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901649 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901659 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901668 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901677 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901687 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901698 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901708 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901717 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901726 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901735 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901745 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901755 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901765 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901778 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901787 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901798 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901808 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901816 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901826 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901835 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901846 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901858 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901868 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901879 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901888 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901904 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901916 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901927 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901939 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901949 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901963 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901975 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901986 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.901996 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902007 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902018 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902028 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902039 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902048 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902057 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902066 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902075 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902086 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902097 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902107 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902116 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902125 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902137 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902146 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902157 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902167 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902179 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902201 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902214 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902225 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902237 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902249 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902259 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902269 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902296 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902306 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902318 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902328 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902339 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902349 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902358 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902368 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902377 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902386 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902397 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902405 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902413 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902425 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902433 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902443 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902453 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902463 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902474 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902484 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902501 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902510 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.902519 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903041 4945 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903063 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903076 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903808 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903824 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903835 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903908 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903922 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903935 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903956 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.903968 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.905651 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.905988 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906001 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906017 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906052 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906064 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906076 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906087 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906099 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906111 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906123 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906137 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906148 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906161 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906176 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906191 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906206 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906223 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906236 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906248 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906261 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906296 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906309 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906321 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906334 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906347 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906359 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906371 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906384 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906397 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906410 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906421 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906434 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906446 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906457 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906471 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906482 4945 reconstruct.go:97] "Volume reconstruction finished" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.906490 4945 reconciler.go:26] "Reconciler: start to sync state" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.922362 4945 manager.go:324] Recovery completed Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.937351 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.943852 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.943907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.943921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.947022 4945 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.947049 4945 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.947075 4945 state_mem.go:36] "Initialized new in-memory state store" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.949477 4945 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.951696 4945 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.951739 4945 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.951992 4945 kubelet.go:2335] "Starting kubelet main sync loop" Dec 06 06:53:06 crc kubenswrapper[4945]: E1206 06:53:06.952062 4945 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 06 06:53:06 crc kubenswrapper[4945]: W1206 06:53:06.952704 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:06 crc kubenswrapper[4945]: E1206 06:53:06.952764 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.220:6443: connect: connection refused" logger="UnhandledError" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.960492 4945 policy_none.go:49] "None policy: Start" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.961692 4945 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 06 06:53:06 crc kubenswrapper[4945]: I1206 06:53:06.961729 4945 state_mem.go:35] "Initializing new in-memory state store" Dec 06 06:53:06 crc kubenswrapper[4945]: E1206 06:53:06.988368 4945 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.021540 4945 manager.go:334] "Starting Device Plugin manager" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.021621 4945 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.021640 4945 server.go:79] "Starting device plugin registration server" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.022132 4945 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.022156 4945 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.022361 4945 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.022516 4945 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.022540 4945 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 06 06:53:07 crc kubenswrapper[4945]: E1206 06:53:07.031550 4945 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.052891 4945 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.053002 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.054020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.054063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.054091 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.054353 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.054616 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.054664 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.055499 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.055542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.055558 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.055731 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.055842 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.055875 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.056040 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.056066 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.056078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.056850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.056879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.056889 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.056969 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057077 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057121 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057718 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057792 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057828 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057945 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.057724 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.058011 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.058810 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.058844 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.058857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.058948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.059014 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.059035 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.059085 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.059112 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.059975 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.060016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.060033 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: E1206 06:53:07.091492 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="400ms" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.108675 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.108778 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.108800 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.108818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.108854 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.108890 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.108930 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.108981 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.109086 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.109151 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.109192 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.109216 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.109243 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.109297 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.109384 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.123051 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.126771 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.126920 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.126931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.127003 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 06:53:07 crc kubenswrapper[4945]: E1206 06:53:07.129071 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.220:6443: connect: connection refused" node="crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210408 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210518 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210571 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210620 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210661 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210703 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210710 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210742 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210747 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210803 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210825 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210874 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210880 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210923 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210942 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210953 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210950 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210959 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211018 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211108 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211011 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211043 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.210880 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211064 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211262 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211341 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211376 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211383 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211562 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.211720 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.330036 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.331699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.331749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.331762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.331794 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 06:53:07 crc kubenswrapper[4945]: E1206 06:53:07.332276 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.220:6443: connect: connection refused" node="crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.384194 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.409566 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.416567 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.434558 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: W1206 06:53:07.437639 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-a44a819ccaa1c62ed3140dee33845f7f9f7edf5235e6caab822835ee091b54e8 WatchSource:0}: Error finding container a44a819ccaa1c62ed3140dee33845f7f9f7edf5235e6caab822835ee091b54e8: Status 404 returned error can't find the container with id a44a819ccaa1c62ed3140dee33845f7f9f7edf5235e6caab822835ee091b54e8 Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.438013 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 06:53:07 crc kubenswrapper[4945]: W1206 06:53:07.462979 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-5a53c4c731aecb323216b542963c2f7cec1df68b8df26b469a6453e0cf19b0dc WatchSource:0}: Error finding container 5a53c4c731aecb323216b542963c2f7cec1df68b8df26b469a6453e0cf19b0dc: Status 404 returned error can't find the container with id 5a53c4c731aecb323216b542963c2f7cec1df68b8df26b469a6453e0cf19b0dc Dec 06 06:53:07 crc kubenswrapper[4945]: W1206 06:53:07.464078 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-fe700c8a7d30a5808ff42e2b7235a2f7355ac3a337f9efd5d9742aac5c13cf5a WatchSource:0}: Error finding container fe700c8a7d30a5808ff42e2b7235a2f7355ac3a337f9efd5d9742aac5c13cf5a: Status 404 returned error can't find the container with id fe700c8a7d30a5808ff42e2b7235a2f7355ac3a337f9efd5d9742aac5c13cf5a Dec 06 06:53:07 crc kubenswrapper[4945]: E1206 06:53:07.492477 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="800ms" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.733231 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.735584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.735622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.735633 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.735660 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 06:53:07 crc kubenswrapper[4945]: E1206 06:53:07.736213 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.220:6443: connect: connection refused" node="crc" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.888027 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:07 crc kubenswrapper[4945]: W1206 06:53:07.905676 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:07 crc kubenswrapper[4945]: E1206 06:53:07.905793 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.220:6443: connect: connection refused" logger="UnhandledError" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.960971 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4"} Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.961158 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a44a819ccaa1c62ed3140dee33845f7f9f7edf5235e6caab822835ee091b54e8"} Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.966948 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333"} Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.967022 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6c3a19bc73837f7d92dfa24643c094bcf2e269c6b07fd89a24f9eec012247cf4"} Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.967203 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.969774 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.969818 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.969830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.972976 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec"} Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.973054 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5a53c4c731aecb323216b542963c2f7cec1df68b8df26b469a6453e0cf19b0dc"} Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.973194 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.974452 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb"} Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.974501 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fe700c8a7d30a5808ff42e2b7235a2f7355ac3a337f9efd5d9742aac5c13cf5a"} Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.974609 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.975581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.975605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.975618 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.975916 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.975950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.975962 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.977538 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac"} Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.977620 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ccaf076f5bedc56719a5be66989bbc17107fe8d3952df16c5de86cc870ac3b7b"} Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.977796 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.978975 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.979024 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:07 crc kubenswrapper[4945]: I1206 06:53:07.979044 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:08 crc kubenswrapper[4945]: W1206 06:53:08.294783 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:08 crc kubenswrapper[4945]: E1206 06:53:08.294911 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.220:6443: connect: connection refused" logger="UnhandledError" Dec 06 06:53:08 crc kubenswrapper[4945]: E1206 06:53:08.294808 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="1.6s" Dec 06 06:53:08 crc kubenswrapper[4945]: W1206 06:53:08.390426 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:08 crc kubenswrapper[4945]: E1206 06:53:08.390527 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.220:6443: connect: connection refused" logger="UnhandledError" Dec 06 06:53:08 crc kubenswrapper[4945]: W1206 06:53:08.482047 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:08 crc kubenswrapper[4945]: E1206 06:53:08.482157 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.220:6443: connect: connection refused" logger="UnhandledError" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.536602 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.539137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.539178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.539188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.539223 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 06:53:08 crc kubenswrapper[4945]: E1206 06:53:08.539838 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.220:6443: connect: connection refused" node="crc" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.886814 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.984420 4945 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333" exitCode=0 Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.984569 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333"} Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.984832 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.986127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.986177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.986187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.986376 4945 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec" exitCode=0 Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.986498 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec"} Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.986643 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.987399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.987423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.987433 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.988026 4945 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb" exitCode=0 Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.988080 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb"} Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.988169 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.988827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.988873 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.988884 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.989235 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac" exitCode=0 Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.989273 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac"} Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.989411 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.990579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.990605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.990619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.994882 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.995511 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337"} Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.995535 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903"} Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.995546 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b"} Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.995664 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.995807 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.995834 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.995845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.997411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.997443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.997457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:08 crc kubenswrapper[4945]: I1206 06:53:08.998340 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:09 crc kubenswrapper[4945]: I1206 06:53:09.886613 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.220:6443: connect: connection refused Dec 06 06:53:09 crc kubenswrapper[4945]: E1206 06:53:09.896147 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="3.2s" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.002203 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4"} Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.002328 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3"} Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.002343 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22"} Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.002514 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.004079 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.004120 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.004132 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.005081 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb"} Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.005111 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2"} Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.005124 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41"} Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.005137 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5"} Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.008079 4945 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36" exitCode=0 Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.008149 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36"} Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.008405 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.009361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.009405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.009418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.010033 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7cc3d8f60c331dcab45117a8f5825236ba0b3b30939fd45727d2c2384d81964b"} Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.010057 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.010268 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.011524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.011562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.011565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.011577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.011588 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.011607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.140873 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.146495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.146557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.146573 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.146610 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 06:53:10 crc kubenswrapper[4945]: E1206 06:53:10.147245 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.220:6443: connect: connection refused" node="crc" Dec 06 06:53:10 crc kubenswrapper[4945]: I1206 06:53:10.671547 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.018257 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770"} Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.018325 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.020147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.020241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.020262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.022500 4945 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a" exitCode=0 Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.022601 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a"} Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.022662 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.022806 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.023899 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.023965 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.023987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.024623 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.024684 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.024707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.173349 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:11 crc kubenswrapper[4945]: I1206 06:53:11.993892 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:12 crc kubenswrapper[4945]: I1206 06:53:12.030931 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4"} Dec 06 06:53:12 crc kubenswrapper[4945]: I1206 06:53:12.031003 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833"} Dec 06 06:53:12 crc kubenswrapper[4945]: I1206 06:53:12.031027 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:12 crc kubenswrapper[4945]: I1206 06:53:12.031124 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:12 crc kubenswrapper[4945]: I1206 06:53:12.032466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:12 crc kubenswrapper[4945]: I1206 06:53:12.032520 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:12 crc kubenswrapper[4945]: I1206 06:53:12.032536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:12 crc kubenswrapper[4945]: I1206 06:53:12.033649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:12 crc kubenswrapper[4945]: I1206 06:53:12.033691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:12 crc kubenswrapper[4945]: I1206 06:53:12.033707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.043260 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676"} Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.043385 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0"} Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.043414 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.043462 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.043417 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67"} Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.045673 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.045713 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.045728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.045985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.046040 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.046069 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.348167 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.350519 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.350596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.350680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.350807 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.638904 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.639619 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.641423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.641474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.641491 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.645233 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.672179 4945 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:53:13 crc kubenswrapper[4945]: I1206 06:53:13.672405 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.046714 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.046767 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.049133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.049162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.049175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.049204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.049241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.049223 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.349518 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.349790 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.352035 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.352114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:14 crc kubenswrapper[4945]: I1206 06:53:14.352140 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:15 crc kubenswrapper[4945]: I1206 06:53:15.183055 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 06 06:53:15 crc kubenswrapper[4945]: I1206 06:53:15.183401 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:15 crc kubenswrapper[4945]: I1206 06:53:15.187497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:15 crc kubenswrapper[4945]: I1206 06:53:15.187747 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:15 crc kubenswrapper[4945]: I1206 06:53:15.187779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:17 crc kubenswrapper[4945]: E1206 06:53:17.031661 4945 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 06:53:17 crc kubenswrapper[4945]: I1206 06:53:17.741407 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 06 06:53:17 crc kubenswrapper[4945]: I1206 06:53:17.741683 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:17 crc kubenswrapper[4945]: I1206 06:53:17.743548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:17 crc kubenswrapper[4945]: I1206 06:53:17.743598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:17 crc kubenswrapper[4945]: I1206 06:53:17.743614 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.829587 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.830013 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.831877 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.831941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.831963 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.976406 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.976680 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.978670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.978968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.978991 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:18 crc kubenswrapper[4945]: I1206 06:53:18.984138 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:19 crc kubenswrapper[4945]: I1206 06:53:19.063267 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:19 crc kubenswrapper[4945]: I1206 06:53:19.065092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:19 crc kubenswrapper[4945]: I1206 06:53:19.065169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:19 crc kubenswrapper[4945]: I1206 06:53:19.065187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:20 crc kubenswrapper[4945]: W1206 06:53:20.474844 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 06 06:53:20 crc kubenswrapper[4945]: I1206 06:53:20.475000 4945 trace.go:236] Trace[191030279]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 06:53:10.472) (total time: 10002ms): Dec 06 06:53:20 crc kubenswrapper[4945]: Trace[191030279]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:53:20.474) Dec 06 06:53:20 crc kubenswrapper[4945]: Trace[191030279]: [10.002081838s] [10.002081838s] END Dec 06 06:53:20 crc kubenswrapper[4945]: E1206 06:53:20.475045 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 06 06:53:20 crc kubenswrapper[4945]: W1206 06:53:20.585124 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 06 06:53:20 crc kubenswrapper[4945]: I1206 06:53:20.585362 4945 trace.go:236] Trace[1507405383]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 06:53:10.583) (total time: 10001ms): Dec 06 06:53:20 crc kubenswrapper[4945]: Trace[1507405383]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:53:20.585) Dec 06 06:53:20 crc kubenswrapper[4945]: Trace[1507405383]: [10.001945254s] [10.001945254s] END Dec 06 06:53:20 crc kubenswrapper[4945]: E1206 06:53:20.585429 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 06 06:53:20 crc kubenswrapper[4945]: I1206 06:53:20.726870 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46984->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 06 06:53:20 crc kubenswrapper[4945]: I1206 06:53:20.726956 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46984->192.168.126.11:17697: read: connection reset by peer" Dec 06 06:53:20 crc kubenswrapper[4945]: W1206 06:53:20.824244 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 06 06:53:20 crc kubenswrapper[4945]: I1206 06:53:20.824390 4945 trace.go:236] Trace[1504357851]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 06:53:10.822) (total time: 10002ms): Dec 06 06:53:20 crc kubenswrapper[4945]: Trace[1504357851]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:53:20.824) Dec 06 06:53:20 crc kubenswrapper[4945]: Trace[1504357851]: [10.002035308s] [10.002035308s] END Dec 06 06:53:20 crc kubenswrapper[4945]: E1206 06:53:20.824422 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 06 06:53:20 crc kubenswrapper[4945]: I1206 06:53:20.886376 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.070564 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.072427 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770" exitCode=255 Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.072505 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770"} Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.072757 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.074006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.074046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.074067 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.074913 4945 scope.go:117] "RemoveContainer" containerID="67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770" Dec 06 06:53:21 crc kubenswrapper[4945]: W1206 06:53:21.295042 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.297176 4945 trace.go:236] Trace[1541624238]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 06:53:11.292) (total time: 10002ms): Dec 06 06:53:21 crc kubenswrapper[4945]: Trace[1541624238]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (06:53:21.295) Dec 06 06:53:21 crc kubenswrapper[4945]: Trace[1541624238]: [10.002229642s] [10.002229642s] END Dec 06 06:53:21 crc kubenswrapper[4945]: E1206 06:53:21.297300 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.664895 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.664979 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.669867 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 06:53:21 crc kubenswrapper[4945]: I1206 06:53:21.669968 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 06:53:22 crc kubenswrapper[4945]: I1206 06:53:22.077899 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 06:53:22 crc kubenswrapper[4945]: I1206 06:53:22.081078 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5"} Dec 06 06:53:22 crc kubenswrapper[4945]: I1206 06:53:22.081257 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:22 crc kubenswrapper[4945]: I1206 06:53:22.082672 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:22 crc kubenswrapper[4945]: I1206 06:53:22.082719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:22 crc kubenswrapper[4945]: I1206 06:53:22.082732 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:23 crc kubenswrapper[4945]: I1206 06:53:23.673252 4945 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:53:23 crc kubenswrapper[4945]: I1206 06:53:23.673377 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 06:53:24 crc kubenswrapper[4945]: I1206 06:53:24.359311 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:24 crc kubenswrapper[4945]: I1206 06:53:24.359551 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:24 crc kubenswrapper[4945]: I1206 06:53:24.359619 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:24 crc kubenswrapper[4945]: I1206 06:53:24.361166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:24 crc kubenswrapper[4945]: I1206 06:53:24.361274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:24 crc kubenswrapper[4945]: I1206 06:53:24.361336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:24 crc kubenswrapper[4945]: I1206 06:53:24.364730 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:24 crc kubenswrapper[4945]: I1206 06:53:24.745071 4945 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 06:53:24 crc kubenswrapper[4945]: I1206 06:53:24.813274 4945 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.093494 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.094572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.094647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.094670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.218439 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.218646 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.220161 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.220209 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.220220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.237436 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.575552 4945 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.888871 4945 apiserver.go:52] "Watching apiserver" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.896477 4945 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.897493 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.898190 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.898456 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.898582 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 06:53:25 crc kubenswrapper[4945]: E1206 06:53:25.898585 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.900052 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:25 crc kubenswrapper[4945]: E1206 06:53:25.900207 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.900372 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.900901 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:25 crc kubenswrapper[4945]: E1206 06:53:25.901258 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.902217 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.902258 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.902433 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.902544 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.902961 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.902967 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.903901 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.903910 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.904004 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.948739 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.962459 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.981406 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:25 crc kubenswrapper[4945]: I1206 06:53:25.989031 4945 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.002584 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.021479 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.036513 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.051031 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.109954 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.213134 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.647585 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.651426 4945 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.651463 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.751850 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.751900 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.751928 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.751951 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.751976 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.752001 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.752026 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.752461 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.752491 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.752460 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.752504 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.752736 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.752949 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.752972 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.753520 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.752056 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.753632 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.753661 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.754353 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.754261 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.754397 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.754431 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.754915 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.754462 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.754997 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.755011 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.755040 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.756185 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.755171 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.755789 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.756104 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.756679 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.756749 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.757070 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.757110 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.757159 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.757191 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.757749 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.757880 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.757189 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.758063 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.758652 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.758744 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.759540 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.759623 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.760061 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.760975 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.760145 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.761062 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.761098 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.761654 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.761708 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.761910 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762040 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762106 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762148 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762183 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762219 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762265 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762423 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762460 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762508 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762548 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762590 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762623 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762634 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762656 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762706 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762714 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762740 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762774 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762799 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762886 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762909 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762932 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762956 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762978 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762999 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763018 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763037 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763055 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763078 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763098 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763115 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763133 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763162 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763180 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763202 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763225 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763243 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763260 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763295 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763321 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763340 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763358 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763377 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763393 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763409 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763489 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763516 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763543 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763565 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763586 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763603 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.762899 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763620 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763050 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763645 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763687 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763710 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763729 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763751 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763771 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763791 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763810 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763827 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763847 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763864 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763884 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763917 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763990 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764015 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764039 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764069 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764091 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764115 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764147 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764225 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764246 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764272 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764325 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764345 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764362 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764382 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764399 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764418 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764440 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764457 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764475 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764494 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764511 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764529 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764545 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764562 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764580 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764599 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764616 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764636 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764663 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764687 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764706 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764725 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764744 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764768 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764787 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764810 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764828 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764846 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764865 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764884 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764900 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764921 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764940 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764957 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764976 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764992 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765010 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765030 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765047 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765063 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765079 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765097 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765112 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765128 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765145 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765163 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765181 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765200 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765217 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765234 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765257 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765274 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765440 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765460 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765476 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765492 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765511 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765530 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765547 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765566 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765654 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765676 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765696 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765715 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765734 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765755 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765775 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765794 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765812 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765830 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765848 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765866 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765886 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765907 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765976 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766011 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766035 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766054 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766071 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766088 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766106 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766126 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766149 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766172 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766191 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766210 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766228 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766245 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766271 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763076 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766296 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763347 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763503 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763595 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763595 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763687 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763691 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763828 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763893 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763932 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763936 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.763982 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764090 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764151 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764200 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764232 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764243 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.764545 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765344 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765562 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765634 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765674 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765839 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765886 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.765959 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766109 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766097 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766126 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766250 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766369 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766304 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766685 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766713 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766755 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766783 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766812 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766819 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766839 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766882 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766919 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766951 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.766958 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767053 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767089 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767098 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.767204 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767225 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767235 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.767303 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:27.267261168 +0000 UTC m=+20.722122422 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767345 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767381 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767417 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767460 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767483 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767488 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767520 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767553 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767582 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767606 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767629 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767651 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767727 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767815 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767837 4945 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767865 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767880 4945 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767891 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767901 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767911 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767921 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767932 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767943 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767954 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767965 4945 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767974 4945 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767984 4945 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767995 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768005 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768014 4945 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768024 4945 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768034 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768042 4945 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768052 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768060 4945 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768070 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768080 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768090 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768114 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768126 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768137 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768147 4945 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768156 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768166 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768177 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768188 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768197 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768208 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768217 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768228 4945 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768239 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768248 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768258 4945 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768269 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768297 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768312 4945 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768324 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768333 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768345 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768355 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768364 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768396 4945 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768408 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768424 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768440 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768454 4945 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768467 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768480 4945 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768493 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768506 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768518 4945 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768531 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768543 4945 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768555 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768567 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768584 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768600 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768613 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768651 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768665 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768677 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768690 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.767519 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.791377 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768243 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768446 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768939 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.768966 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.791523 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.791784 4945 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.792462 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.792672 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.792796 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.792983 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.769418 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.769613 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.769845 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.770164 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.770191 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.770804 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.770879 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.770904 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.771100 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.771100 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.771166 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.771578 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.771723 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.771808 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.771836 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.771990 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.772112 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.772340 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.772583 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.772788 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.773223 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.797982 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.773333 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.773728 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.773849 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.773907 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.774097 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.774340 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.798051 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.775264 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.775555 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.775629 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.775706 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.775963 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.776044 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.776109 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.776495 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.781528 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.782041 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.782914 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.782969 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.783274 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.784211 4945 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.798308 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.784400 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.784619 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.785530 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.798391 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.798410 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.786113 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.786997 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.787239 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.787246 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.787425 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.787449 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.787733 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.787778 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.788519 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.788528 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.788677 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.788865 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.788926 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.788995 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.789196 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.789382 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.789585 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.789759 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.789794 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.790153 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.790171 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.790350 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.790507 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.790708 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.790759 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.790883 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.791330 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.791371 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.793851 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.794176 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.794352 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.794668 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.794798 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.795337 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.795856 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.793898 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.796549 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.796565 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.796589 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.797327 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.797503 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:27.297481091 +0000 UTC m=+20.752342135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.797704 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.797786 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.774914 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.798621 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.798798 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.798872 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.799157 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:27.29908775 +0000 UTC m=+20.753948794 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.799243 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.799347 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.799389 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.799403 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.799422 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.799461 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.799656 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.799698 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.799823 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:53:27.299805047 +0000 UTC m=+20.754666291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.800137 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.800189 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.801464 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.801499 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.805377 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.805589 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.805926 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.806065 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.806137 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:26 crc kubenswrapper[4945]: E1206 06:53:26.806252 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:27.30622749 +0000 UTC m=+20.761088534 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.806052 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.806939 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.805971 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.806543 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.806617 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.807189 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.807588 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.807673 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.807827 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.809338 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.814887 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.815510 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.820138 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.828766 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.836917 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.839458 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.841476 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 06:53:26 crc kubenswrapper[4945]: W1206 06:53:26.860966 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-1a1806022f5c6beedcbe30ea63a3f895020061d23b7e76ff72cb94c3d0cc25c0 WatchSource:0}: Error finding container 1a1806022f5c6beedcbe30ea63a3f895020061d23b7e76ff72cb94c3d0cc25c0: Status 404 returned error can't find the container with id 1a1806022f5c6beedcbe30ea63a3f895020061d23b7e76ff72cb94c3d0cc25c0 Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870187 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870238 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870351 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870368 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870363 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870382 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870445 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870456 4945 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870470 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870481 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870493 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870503 4945 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870512 4945 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870522 4945 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870532 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870544 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870555 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870566 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870575 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870584 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870594 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870603 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870612 4945 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870621 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870631 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870639 4945 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870652 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870662 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870670 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870679 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870689 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870697 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870706 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870715 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870724 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870621 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870736 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870826 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870861 4945 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870881 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870896 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870909 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870922 4945 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870935 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870947 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870960 4945 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870974 4945 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.870987 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871000 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871013 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871026 4945 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871039 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871052 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871082 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871095 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871108 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871121 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871135 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871148 4945 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871163 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871177 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871189 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871201 4945 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871215 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871228 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871241 4945 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871257 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871272 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871315 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871331 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871344 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871357 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871371 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871384 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871399 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871412 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871425 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871438 4945 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871451 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871463 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871475 4945 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871487 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871500 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871513 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871529 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871542 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871554 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871567 4945 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871580 4945 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871593 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871605 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871618 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871655 4945 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871668 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871682 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871694 4945 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871706 4945 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871718 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871731 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871743 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871755 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871767 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871780 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871793 4945 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871807 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871819 4945 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871831 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871845 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871858 4945 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871871 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871883 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871897 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871909 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871921 4945 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871933 4945 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871945 4945 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871957 4945 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871968 4945 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871981 4945 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.871993 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.872008 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.872022 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.872036 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.872049 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.872062 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.872075 4945 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.872089 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.872102 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.872115 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.962176 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.962951 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.965065 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.966642 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.967746 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.968320 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.968870 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.969812 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.970421 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.971325 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.971833 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.972861 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.974303 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.974864 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.975405 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.975915 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.976545 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.976926 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.978425 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.978985 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.979434 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.980444 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.980931 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.981993 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.982212 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.982428 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.983563 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.985821 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.986461 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.987191 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.988063 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.988625 4945 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.988727 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.991635 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.992259 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.992969 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.995796 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.996742 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.997900 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.998673 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.998950 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:26 crc kubenswrapper[4945]: I1206 06:53:26.999903 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.000426 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.001039 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.002066 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.003101 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.003590 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.004524 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.005026 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.006183 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.006726 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.007597 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.008065 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.008619 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.010882 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.011408 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.013848 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.037476 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.055249 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.068574 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.089375 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.101179 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.106008 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839"} Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.106261 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1a1806022f5c6beedcbe30ea63a3f895020061d23b7e76ff72cb94c3d0cc25c0"} Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.128024 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 06:53:27 crc kubenswrapper[4945]: W1206 06:53:27.140512 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-880c5150e5494aa28ab5a0ca113164e8b54b3ad9c3774c253c18e810b28bf9e4 WatchSource:0}: Error finding container 880c5150e5494aa28ab5a0ca113164e8b54b3ad9c3774c253c18e810b28bf9e4: Status 404 returned error can't find the container with id 880c5150e5494aa28ab5a0ca113164e8b54b3ad9c3774c253c18e810b28bf9e4 Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.153988 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 06:53:27 crc kubenswrapper[4945]: W1206 06:53:27.180053 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-2a5d013323278d1ef900ed04e46c6c0f0615eb7a8773019d8e869df49f460f95 WatchSource:0}: Error finding container 2a5d013323278d1ef900ed04e46c6c0f0615eb7a8773019d8e869df49f460f95: Status 404 returned error can't find the container with id 2a5d013323278d1ef900ed04e46c6c0f0615eb7a8773019d8e869df49f460f95 Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.275475 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.275597 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.275654 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:28.275639839 +0000 UTC m=+21.730500883 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.376911 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.377067 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.377109 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377174 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:53:28.377134637 +0000 UTC m=+21.831995681 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377237 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.377273 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377270 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377350 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:28.377328032 +0000 UTC m=+21.832189076 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377359 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377371 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377383 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377392 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377402 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377441 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:28.377434734 +0000 UTC m=+21.832295778 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.377476 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:28.377453655 +0000 UTC m=+21.832314889 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.952976 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.953056 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.953143 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.953303 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:27 crc kubenswrapper[4945]: I1206 06:53:27.954591 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:27 crc kubenswrapper[4945]: E1206 06:53:27.954963 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.110108 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260"} Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.110155 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"880c5150e5494aa28ab5a0ca113164e8b54b3ad9c3774c253c18e810b28bf9e4"} Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.112942 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993"} Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.114270 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2a5d013323278d1ef900ed04e46c6c0f0615eb7a8773019d8e869df49f460f95"} Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.135786 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.152865 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.167803 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.190524 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.203863 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.221636 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.239959 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.254735 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.270827 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.286410 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.287202 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.287318 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:30.287270269 +0000 UTC m=+23.742131413 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.296715 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.316389 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.380080 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.386820 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.386921 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.386953 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.386978 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:53:30.386949933 +0000 UTC m=+23.841810977 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.387023 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.387054 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.387094 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:30.387086326 +0000 UTC m=+23.841947370 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.387169 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.387187 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.387201 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.387228 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:30.38722065 +0000 UTC m=+23.842081694 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.387170 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.387248 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.387256 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:28 crc kubenswrapper[4945]: E1206 06:53:28.387306 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:30.387296732 +0000 UTC m=+23.842157786 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.408041 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.437729 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.462954 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.493208 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.727052 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-hw42h"] Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.727587 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hw42h" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.732038 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.732308 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.732375 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.749933 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.764712 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.777859 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.803155 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.818674 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.834122 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.851083 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.865697 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.885268 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:28Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.891873 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d60d17d4-5e9c-4e43-b584-54e7187712b7-hosts-file\") pod \"node-resolver-hw42h\" (UID: \"d60d17d4-5e9c-4e43-b584-54e7187712b7\") " pod="openshift-dns/node-resolver-hw42h" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.891959 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlvjz\" (UniqueName: \"kubernetes.io/projected/d60d17d4-5e9c-4e43-b584-54e7187712b7-kube-api-access-vlvjz\") pod \"node-resolver-hw42h\" (UID: \"d60d17d4-5e9c-4e43-b584-54e7187712b7\") " pod="openshift-dns/node-resolver-hw42h" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.993112 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d60d17d4-5e9c-4e43-b584-54e7187712b7-hosts-file\") pod \"node-resolver-hw42h\" (UID: \"d60d17d4-5e9c-4e43-b584-54e7187712b7\") " pod="openshift-dns/node-resolver-hw42h" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.993167 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlvjz\" (UniqueName: \"kubernetes.io/projected/d60d17d4-5e9c-4e43-b584-54e7187712b7-kube-api-access-vlvjz\") pod \"node-resolver-hw42h\" (UID: \"d60d17d4-5e9c-4e43-b584-54e7187712b7\") " pod="openshift-dns/node-resolver-hw42h" Dec 06 06:53:28 crc kubenswrapper[4945]: I1206 06:53:28.993488 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d60d17d4-5e9c-4e43-b584-54e7187712b7-hosts-file\") pod \"node-resolver-hw42h\" (UID: \"d60d17d4-5e9c-4e43-b584-54e7187712b7\") " pod="openshift-dns/node-resolver-hw42h" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.017118 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlvjz\" (UniqueName: \"kubernetes.io/projected/d60d17d4-5e9c-4e43-b584-54e7187712b7-kube-api-access-vlvjz\") pod \"node-resolver-hw42h\" (UID: \"d60d17d4-5e9c-4e43-b584-54e7187712b7\") " pod="openshift-dns/node-resolver-hw42h" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.041653 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hw42h" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.119008 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hw42h" event={"ID":"d60d17d4-5e9c-4e43-b584-54e7187712b7","Type":"ContainerStarted","Data":"d8674db04dfe03003571a71ef45d56bef27f147866d0535972a19fa482cc4268"} Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.124502 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-6v6gj"] Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.124964 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-jrv66"] Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.124982 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.126334 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-57tvp"] Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.126396 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.127008 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.127759 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.127801 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.127841 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mlqp5"] Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.127950 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.128975 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.129563 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.129814 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.130009 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.130193 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.130356 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.130446 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.132567 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.143554 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.144018 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.145049 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.157762 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.157890 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.157798 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.158257 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.158300 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.158262 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.195229 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.225057 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.249294 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295311 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295385 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grvtv\" (UniqueName: \"kubernetes.io/projected/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-kube-api-access-grvtv\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295423 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-etc-kubernetes\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295442 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-kubelet\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295459 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-conf-dir\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295479 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-var-lib-cni-bin\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295494 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-config\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295514 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qh8t\" (UniqueName: \"kubernetes.io/projected/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-kube-api-access-5qh8t\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295544 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-script-lib\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295562 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-socket-dir-parent\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-run-multus-certs\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295595 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295612 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-netns\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295627 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-os-release\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295650 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-cnibin\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295801 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295874 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-cni-binary-copy\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295902 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f56279eb-fe4e-4bd2-bccd-d39958e923ea-rootfs\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295927 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-etc-openvswitch\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295944 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-bin\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295961 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-cnibin\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295980 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-openvswitch\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.295996 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-ovn\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296011 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-env-overrides\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296028 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-var-lib-cni-multus\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296059 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f379f13c-33b1-4e56-81d5-9244ceb6cabd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296074 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-systemd\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296089 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-daemon-config\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296105 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-ovn-kubernetes\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296124 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-cni-dir\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296173 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g659f\" (UniqueName: \"kubernetes.io/projected/f379f13c-33b1-4e56-81d5-9244ceb6cabd-kube-api-access-g659f\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296197 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-slash\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296238 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-systemd-units\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296258 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovn-node-metrics-cert\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296295 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-run-k8s-cni-cncf-io\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296345 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f56279eb-fe4e-4bd2-bccd-d39958e923ea-mcd-auth-proxy-config\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296369 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-os-release\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296391 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-node-log\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296411 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-netd\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296429 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-system-cni-dir\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-var-lib-openvswitch\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296480 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-system-cni-dir\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296497 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hhpb\" (UniqueName: \"kubernetes.io/projected/f56279eb-fe4e-4bd2-bccd-d39958e923ea-kube-api-access-7hhpb\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296513 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-run-netns\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296530 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-var-lib-kubelet\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296545 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-hostroot\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296562 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f56279eb-fe4e-4bd2-bccd-d39958e923ea-proxy-tls\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296591 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f379f13c-33b1-4e56-81d5-9244ceb6cabd-cni-binary-copy\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.296615 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-log-socket\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.325474 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.344239 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.373200 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.391922 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.397968 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-var-lib-openvswitch\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398004 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-system-cni-dir\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398027 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hhpb\" (UniqueName: \"kubernetes.io/projected/f56279eb-fe4e-4bd2-bccd-d39958e923ea-kube-api-access-7hhpb\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398081 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f379f13c-33b1-4e56-81d5-9244ceb6cabd-cni-binary-copy\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398099 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-log-socket\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398120 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-run-netns\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398136 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-var-lib-kubelet\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-hostroot\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398140 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-var-lib-openvswitch\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398220 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-system-cni-dir\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398168 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f56279eb-fe4e-4bd2-bccd-d39958e923ea-proxy-tls\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398795 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grvtv\" (UniqueName: \"kubernetes.io/projected/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-kube-api-access-grvtv\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398843 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-kubelet\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398864 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-etc-kubernetes\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398892 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-var-lib-cni-bin\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398894 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-run-netns\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398908 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-conf-dir\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398909 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-var-lib-kubelet\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398930 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-config\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398949 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qh8t\" (UniqueName: \"kubernetes.io/projected/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-kube-api-access-5qh8t\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398956 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-log-socket\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398960 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-etc-kubernetes\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398966 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-script-lib\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.398989 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-socket-dir-parent\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399009 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-run-multus-certs\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399011 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-var-lib-cni-bin\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399023 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-conf-dir\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399057 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-cnibin\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399070 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-hostroot\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399088 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-cnibin\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399009 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-kubelet\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399089 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-socket-dir-parent\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399089 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399143 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f379f13c-33b1-4e56-81d5-9244ceb6cabd-cni-binary-copy\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399171 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-netns\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399180 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-run-multus-certs\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399199 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-os-release\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399220 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-netns\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399231 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399261 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399262 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-etc-openvswitch\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399308 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-etc-openvswitch\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399319 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-bin\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399342 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-cnibin\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399361 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-cni-binary-copy\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399380 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f56279eb-fe4e-4bd2-bccd-d39958e923ea-rootfs\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399403 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f379f13c-33b1-4e56-81d5-9244ceb6cabd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399343 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-bin\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399422 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-systemd\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399443 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-openvswitch\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399461 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-ovn\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399480 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-env-overrides\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399497 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-var-lib-cni-multus\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399515 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-daemon-config\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399520 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-os-release\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399537 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-ovn-kubernetes\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399556 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-cni-dir\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399561 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f56279eb-fe4e-4bd2-bccd-d39958e923ea-rootfs\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399572 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g659f\" (UniqueName: \"kubernetes.io/projected/f379f13c-33b1-4e56-81d5-9244ceb6cabd-kube-api-access-g659f\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399589 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-slash\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399623 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-systemd-units\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399640 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovn-node-metrics-cert\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399656 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-run-k8s-cni-cncf-io\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399671 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-node-log\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399688 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-netd\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399708 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-system-cni-dir\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399726 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f56279eb-fe4e-4bd2-bccd-d39958e923ea-mcd-auth-proxy-config\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399743 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-os-release\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399812 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-config\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399858 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-openvswitch\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399868 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-os-release\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399885 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-systemd\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399898 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-var-lib-cni-multus\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399914 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-systemd-units\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400032 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f379f13c-33b1-4e56-81d5-9244ceb6cabd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400076 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-ovn\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400104 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-env-overrides\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.399461 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-cnibin\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400152 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-script-lib\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400256 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-slash\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400305 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-cni-dir\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400335 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-ovn-kubernetes\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400341 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-netd\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400366 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-host-run-k8s-cni-cncf-io\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400373 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-system-cni-dir\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400401 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-node-log\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400396 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-cni-binary-copy\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400448 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-multus-daemon-config\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.400858 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f56279eb-fe4e-4bd2-bccd-d39958e923ea-mcd-auth-proxy-config\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.401673 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f379f13c-33b1-4e56-81d5-9244ceb6cabd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.403901 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f56279eb-fe4e-4bd2-bccd-d39958e923ea-proxy-tls\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.403915 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovn-node-metrics-cert\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.419215 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.424508 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hhpb\" (UniqueName: \"kubernetes.io/projected/f56279eb-fe4e-4bd2-bccd-d39958e923ea-kube-api-access-7hhpb\") pod \"machine-config-daemon-57tvp\" (UID: \"f56279eb-fe4e-4bd2-bccd-d39958e923ea\") " pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.431629 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g659f\" (UniqueName: \"kubernetes.io/projected/f379f13c-33b1-4e56-81d5-9244ceb6cabd-kube-api-access-g659f\") pod \"multus-additional-cni-plugins-jrv66\" (UID: \"f379f13c-33b1-4e56-81d5-9244ceb6cabd\") " pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.431663 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qh8t\" (UniqueName: \"kubernetes.io/projected/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-kube-api-access-5qh8t\") pod \"ovnkube-node-mlqp5\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.432044 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grvtv\" (UniqueName: \"kubernetes.io/projected/c9fc852c-2d5d-444c-b470-2728bb3dcdfd-kube-api-access-grvtv\") pod \"multus-6v6gj\" (UID: \"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\") " pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.432761 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.444337 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6v6gj" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.446038 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.454243 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jrv66" Dec 06 06:53:29 crc kubenswrapper[4945]: W1206 06:53:29.459984 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9fc852c_2d5d_444c_b470_2728bb3dcdfd.slice/crio-7ff2dece9cdd9ef4d1cdfb0b868516b156498732939cb934a3722be40ae219bc WatchSource:0}: Error finding container 7ff2dece9cdd9ef4d1cdfb0b868516b156498732939cb934a3722be40ae219bc: Status 404 returned error can't find the container with id 7ff2dece9cdd9ef4d1cdfb0b868516b156498732939cb934a3722be40ae219bc Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.461397 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.466722 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:53:29 crc kubenswrapper[4945]: W1206 06:53:29.469334 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf379f13c_33b1_4e56_81d5_9244ceb6cabd.slice/crio-2cbfb1e3659dde278e4baa267e3a8f72c43d061736dce845b58d1b3b51919f3d WatchSource:0}: Error finding container 2cbfb1e3659dde278e4baa267e3a8f72c43d061736dce845b58d1b3b51919f3d: Status 404 returned error can't find the container with id 2cbfb1e3659dde278e4baa267e3a8f72c43d061736dce845b58d1b3b51919f3d Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.479377 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: W1206 06:53:29.490801 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf56279eb_fe4e_4bd2_bccd_d39958e923ea.slice/crio-3713fbf8981a23479834272cdff04c6aefa3486ea18a21dda04b5ff5b987768b WatchSource:0}: Error finding container 3713fbf8981a23479834272cdff04c6aefa3486ea18a21dda04b5ff5b987768b: Status 404 returned error can't find the container with id 3713fbf8981a23479834272cdff04c6aefa3486ea18a21dda04b5ff5b987768b Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.491041 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.497194 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.513008 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.528012 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.551649 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.566453 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.580832 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.597074 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.609554 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.626243 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.650154 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.953630 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.953782 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:29 crc kubenswrapper[4945]: E1206 06:53:29.953867 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:29 crc kubenswrapper[4945]: I1206 06:53:29.953907 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:29 crc kubenswrapper[4945]: E1206 06:53:29.954044 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:29 crc kubenswrapper[4945]: E1206 06:53:29.954141 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.125105 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6v6gj" event={"ID":"c9fc852c-2d5d-444c-b470-2728bb3dcdfd","Type":"ContainerStarted","Data":"844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.125710 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6v6gj" event={"ID":"c9fc852c-2d5d-444c-b470-2728bb3dcdfd","Type":"ContainerStarted","Data":"7ff2dece9cdd9ef4d1cdfb0b868516b156498732939cb934a3722be40ae219bc"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.127483 4945 generic.go:334] "Generic (PLEG): container finished" podID="f379f13c-33b1-4e56-81d5-9244ceb6cabd" containerID="1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00" exitCode=0 Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.127615 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" event={"ID":"f379f13c-33b1-4e56-81d5-9244ceb6cabd","Type":"ContainerDied","Data":"1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.127716 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" event={"ID":"f379f13c-33b1-4e56-81d5-9244ceb6cabd","Type":"ContainerStarted","Data":"2cbfb1e3659dde278e4baa267e3a8f72c43d061736dce845b58d1b3b51919f3d"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.129314 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.131814 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d" exitCode=0 Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.131851 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.131893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"eba08d28493cc703b338f067dae3837d840455078335644aa955264435c755db"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.134163 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hw42h" event={"ID":"d60d17d4-5e9c-4e43-b584-54e7187712b7","Type":"ContainerStarted","Data":"336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.137184 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.137252 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.137269 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"3713fbf8981a23479834272cdff04c6aefa3486ea18a21dda04b5ff5b987768b"} Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.148507 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.161568 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.174137 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.187975 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.199147 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.221638 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.238068 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.258811 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.274366 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.288048 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.303120 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.309938 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.311983 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.312033 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:34.312017992 +0000 UTC m=+27.766879026 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.320610 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.342572 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.359820 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.374575 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.389550 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.404851 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.410603 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.410717 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.410754 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.410782 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.410896 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.410907 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:53:34.410883767 +0000 UTC m=+27.865744811 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.410902 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.410942 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.410952 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.410968 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.410976 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.410959 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.410948 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:34.410936758 +0000 UTC m=+27.865797802 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.411039 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:34.411027531 +0000 UTC m=+27.865888575 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:30 crc kubenswrapper[4945]: E1206 06:53:30.411060 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:34.411052531 +0000 UTC m=+27.865913575 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.418893 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.432075 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.443573 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.457924 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.478751 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.491608 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.511993 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.528095 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.550639 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.650441 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-np9zn"] Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.650951 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.653809 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.654091 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.654160 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.657167 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.691234 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.695952 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.706786 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.712994 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.730777 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.768528 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.788347 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.805332 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.815345 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d67cbd5-c8f0-499f-9255-01ad8dc4071f-host\") pod \"node-ca-np9zn\" (UID: \"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\") " pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.815393 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkxsj\" (UniqueName: \"kubernetes.io/projected/6d67cbd5-c8f0-499f-9255-01ad8dc4071f-kube-api-access-kkxsj\") pod \"node-ca-np9zn\" (UID: \"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\") " pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.815422 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6d67cbd5-c8f0-499f-9255-01ad8dc4071f-serviceca\") pod \"node-ca-np9zn\" (UID: \"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\") " pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.822638 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.843200 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.864858 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.882070 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.898749 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.910808 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.916019 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d67cbd5-c8f0-499f-9255-01ad8dc4071f-host\") pod \"node-ca-np9zn\" (UID: \"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\") " pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.916064 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkxsj\" (UniqueName: \"kubernetes.io/projected/6d67cbd5-c8f0-499f-9255-01ad8dc4071f-kube-api-access-kkxsj\") pod \"node-ca-np9zn\" (UID: \"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\") " pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.916094 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6d67cbd5-c8f0-499f-9255-01ad8dc4071f-serviceca\") pod \"node-ca-np9zn\" (UID: \"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\") " pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.916207 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d67cbd5-c8f0-499f-9255-01ad8dc4071f-host\") pod \"node-ca-np9zn\" (UID: \"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\") " pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.917269 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6d67cbd5-c8f0-499f-9255-01ad8dc4071f-serviceca\") pod \"node-ca-np9zn\" (UID: \"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\") " pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.926987 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.939004 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkxsj\" (UniqueName: \"kubernetes.io/projected/6d67cbd5-c8f0-499f-9255-01ad8dc4071f-kube-api-access-kkxsj\") pod \"node-ca-np9zn\" (UID: \"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\") " pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.940367 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.953897 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.968297 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.968370 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-np9zn" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.983366 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:30 crc kubenswrapper[4945]: I1206 06:53:30.995454 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.010472 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.028567 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: W1206 06:53:31.040772 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d67cbd5_c8f0_499f_9255_01ad8dc4071f.slice/crio-98e6c5b979826cf0cb635d0d5e6be6adf323c3fbb7b44c285541447edfb41c63 WatchSource:0}: Error finding container 98e6c5b979826cf0cb635d0d5e6be6adf323c3fbb7b44c285541447edfb41c63: Status 404 returned error can't find the container with id 98e6c5b979826cf0cb635d0d5e6be6adf323c3fbb7b44c285541447edfb41c63 Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.049886 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.077744 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.093189 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.107550 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.122735 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.138505 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.142216 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-np9zn" event={"ID":"6d67cbd5-c8f0-499f-9255-01ad8dc4071f","Type":"ContainerStarted","Data":"98e6c5b979826cf0cb635d0d5e6be6adf323c3fbb7b44c285541447edfb41c63"} Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.146821 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059"} Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.146865 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af"} Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.146875 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee"} Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.146884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c"} Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.146893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6"} Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.146902 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259"} Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.149141 4945 generic.go:334] "Generic (PLEG): container finished" podID="f379f13c-33b1-4e56-81d5-9244ceb6cabd" containerID="6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120" exitCode=0 Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.149851 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" event={"ID":"f379f13c-33b1-4e56-81d5-9244ceb6cabd","Type":"ContainerDied","Data":"6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120"} Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.151895 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.166094 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.179623 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.182760 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.203648 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.221446 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.236466 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.249419 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.263866 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.282352 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.297261 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.310999 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.333578 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.347547 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.359759 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.372723 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.386429 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.420967 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.464010 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.507384 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:31Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.952658 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.952681 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:31 crc kubenswrapper[4945]: E1206 06:53:31.953349 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:31 crc kubenswrapper[4945]: I1206 06:53:31.952711 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:31 crc kubenswrapper[4945]: E1206 06:53:31.953601 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:31 crc kubenswrapper[4945]: E1206 06:53:31.953776 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.153560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-np9zn" event={"ID":"6d67cbd5-c8f0-499f-9255-01ad8dc4071f","Type":"ContainerStarted","Data":"a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303"} Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.157079 4945 generic.go:334] "Generic (PLEG): container finished" podID="f379f13c-33b1-4e56-81d5-9244ceb6cabd" containerID="e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2" exitCode=0 Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.157127 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" event={"ID":"f379f13c-33b1-4e56-81d5-9244ceb6cabd","Type":"ContainerDied","Data":"e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2"} Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.178475 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.212074 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.235795 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.257053 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.270669 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.284251 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.299978 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.312267 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.326041 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.349889 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.365029 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.378525 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.394360 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.413887 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.427185 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.440132 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.453151 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.465515 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.480427 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.493688 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.506153 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.518854 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.530419 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.550428 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.565997 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.579271 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.598259 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.627915 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.661423 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:32 crc kubenswrapper[4945]: I1206 06:53:32.702185 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:32Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.051977 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.054460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.054511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.054524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.054600 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.064180 4945 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.064589 4945 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.066003 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.066037 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.066048 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.066074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.066089 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: E1206 06:53:33.078713 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.083023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.083074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.083089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.083112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.083127 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: E1206 06:53:33.095142 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.099522 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.099581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.099595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.099618 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.099641 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: E1206 06:53:33.115300 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.119148 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.119214 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.119227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.119249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.119263 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: E1206 06:53:33.131251 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.134929 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.134973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.134985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.135005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.135019 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: E1206 06:53:33.147863 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: E1206 06:53:33.148000 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.149861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.149913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.149923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.149944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.149956 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.161738 4945 generic.go:334] "Generic (PLEG): container finished" podID="f379f13c-33b1-4e56-81d5-9244ceb6cabd" containerID="cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75" exitCode=0 Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.161808 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" event={"ID":"f379f13c-33b1-4e56-81d5-9244ceb6cabd","Type":"ContainerDied","Data":"cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75"} Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.176588 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.188662 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.201641 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.231766 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.247119 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.253380 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.253413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.253422 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.253436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.253448 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.261118 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.275778 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.298931 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.314109 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.327386 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.340627 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.351321 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.355457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.355494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.355502 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.355518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.355528 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.366786 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.380380 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.394861 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:33Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.458326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.458374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.458384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.458402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.458415 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.562028 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.562086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.562097 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.562115 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.562124 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.666987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.667597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.667835 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.668021 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.668184 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.772858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.772917 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.772930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.772949 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.772965 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.875809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.875851 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.875862 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.875879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.875892 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.953012 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:33 crc kubenswrapper[4945]: E1206 06:53:33.953181 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.953609 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:33 crc kubenswrapper[4945]: E1206 06:53:33.953668 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.953712 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:33 crc kubenswrapper[4945]: E1206 06:53:33.953750 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.979076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.979107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.979118 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.979131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:33 crc kubenswrapper[4945]: I1206 06:53:33.979141 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:33Z","lastTransitionTime":"2025-12-06T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.086799 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.086890 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.086911 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.086930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.086946 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:34Z","lastTransitionTime":"2025-12-06T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.170435 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.173906 4945 generic.go:334] "Generic (PLEG): container finished" podID="f379f13c-33b1-4e56-81d5-9244ceb6cabd" containerID="9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2" exitCode=0 Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.173951 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" event={"ID":"f379f13c-33b1-4e56-81d5-9244ceb6cabd","Type":"ContainerDied","Data":"9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.190935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.191012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.191032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.191063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.191084 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:34Z","lastTransitionTime":"2025-12-06T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.193204 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.207228 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.218607 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.231967 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.255932 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.271878 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.289137 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.294272 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.294347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.294367 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.294390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.294406 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:34Z","lastTransitionTime":"2025-12-06T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.309979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.325699 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.340659 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.357114 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.361016 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.361354 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.361415 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:42.361388467 +0000 UTC m=+35.816249511 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.373118 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.388395 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.397802 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.397852 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.397863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.397882 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.397895 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:34Z","lastTransitionTime":"2025-12-06T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.405219 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.420480 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.462010 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.462182 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462298 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:53:42.46223963 +0000 UTC m=+35.917100674 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462318 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462393 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:42.462370793 +0000 UTC m=+35.917231827 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.462417 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.462450 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462542 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462561 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462575 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462606 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:42.462597388 +0000 UTC m=+35.917458432 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462699 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462746 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462763 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:34 crc kubenswrapper[4945]: E1206 06:53:34.462841 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:42.462816894 +0000 UTC m=+35.917677948 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.501498 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.501557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.501572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.501593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.501669 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:34Z","lastTransitionTime":"2025-12-06T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.605583 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.605661 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.605686 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.605715 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.605737 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:34Z","lastTransitionTime":"2025-12-06T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.707874 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.707923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.707934 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.707953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.707966 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:34Z","lastTransitionTime":"2025-12-06T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.810987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.811041 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.811052 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.811070 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.811081 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:34Z","lastTransitionTime":"2025-12-06T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.914546 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.914605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.914616 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.914635 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:34 crc kubenswrapper[4945]: I1206 06:53:34.914646 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:34Z","lastTransitionTime":"2025-12-06T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.018030 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.018092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.018106 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.018126 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.018140 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:35Z","lastTransitionTime":"2025-12-06T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.121594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.121683 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.121706 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.121736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.121754 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:35Z","lastTransitionTime":"2025-12-06T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.182078 4945 generic.go:334] "Generic (PLEG): container finished" podID="f379f13c-33b1-4e56-81d5-9244ceb6cabd" containerID="7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459" exitCode=0 Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.182144 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" event={"ID":"f379f13c-33b1-4e56-81d5-9244ceb6cabd","Type":"ContainerDied","Data":"7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.198598 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.222864 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.225303 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.225364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.225376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.225397 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.225407 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:35Z","lastTransitionTime":"2025-12-06T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.244577 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.257733 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.275854 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.296864 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.313446 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.328389 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.328445 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.328458 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.328484 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.328498 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:35Z","lastTransitionTime":"2025-12-06T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.330874 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.345269 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.360188 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.373164 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.387552 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.401338 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.416121 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.430640 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.432269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.432324 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.432351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.432372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.432389 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:35Z","lastTransitionTime":"2025-12-06T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.534935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.535377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.535467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.535563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.535645 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:35Z","lastTransitionTime":"2025-12-06T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.639245 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.639323 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.639338 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.639361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.639376 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:35Z","lastTransitionTime":"2025-12-06T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.742528 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.742593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.742605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.742623 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.742634 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:35Z","lastTransitionTime":"2025-12-06T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.845224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.845327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.845347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.845376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.845394 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:35Z","lastTransitionTime":"2025-12-06T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.947827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.947870 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.947883 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.947899 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.947909 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:35Z","lastTransitionTime":"2025-12-06T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.953175 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:35 crc kubenswrapper[4945]: E1206 06:53:35.953389 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.953468 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:35 crc kubenswrapper[4945]: E1206 06:53:35.953603 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:35 crc kubenswrapper[4945]: I1206 06:53:35.953652 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:35 crc kubenswrapper[4945]: E1206 06:53:35.953701 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.050418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.050464 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.050474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.050490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.050500 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:36Z","lastTransitionTime":"2025-12-06T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.152861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.152903 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.152913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.152935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.152946 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:36Z","lastTransitionTime":"2025-12-06T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.191546 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.192132 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.196836 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" event={"ID":"f379f13c-33b1-4e56-81d5-9244ceb6cabd","Type":"ContainerStarted","Data":"21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.206014 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.221917 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.235805 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.249164 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.255744 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.255791 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.255823 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.255842 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.255854 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:36Z","lastTransitionTime":"2025-12-06T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.261140 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.275545 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.283355 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.289989 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.305217 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.320033 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.333624 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.348392 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.358092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.358142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.358153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.358171 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.358181 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:36Z","lastTransitionTime":"2025-12-06T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.370179 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.387149 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.406427 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.424342 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.441178 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.458119 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.464154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.464214 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.464227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.464244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.464258 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:36Z","lastTransitionTime":"2025-12-06T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.485938 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.503025 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.516464 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.527943 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.542947 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.556852 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.567511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.567556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.567569 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.567593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.567608 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:36Z","lastTransitionTime":"2025-12-06T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.570800 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.593713 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.608483 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.625694 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.641853 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.656934 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.669175 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.671205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.671249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.671262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.671317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.671336 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:36Z","lastTransitionTime":"2025-12-06T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.773572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.773624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.773642 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.773666 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.773681 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:36Z","lastTransitionTime":"2025-12-06T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.877595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.877668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.877690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.877718 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.877735 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:36Z","lastTransitionTime":"2025-12-06T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.970460 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.980886 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.980961 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.980976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.981001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.981026 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:36Z","lastTransitionTime":"2025-12-06T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:36 crc kubenswrapper[4945]: I1206 06:53:36.992194 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.013581 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.029354 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.052007 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.068585 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.083625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.083668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.083680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.083702 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.083715 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:37Z","lastTransitionTime":"2025-12-06T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.083896 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.095063 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.106428 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.122097 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.143000 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.156480 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.173000 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.186227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.186299 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.186312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.186328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.186338 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:37Z","lastTransitionTime":"2025-12-06T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.200210 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.200747 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.205456 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.222796 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.225712 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.239877 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.254475 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.269230 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.285415 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.290326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.290364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.290374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.290392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.290403 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:37Z","lastTransitionTime":"2025-12-06T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.299852 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.315481 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.340828 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.357126 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.372387 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.384404 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.392893 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.392945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.392958 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.392976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.392990 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:37Z","lastTransitionTime":"2025-12-06T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.396295 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.407373 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.420784 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.437018 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.458379 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.495581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.495664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.495681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.495703 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.495717 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:37Z","lastTransitionTime":"2025-12-06T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.599210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.599259 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.599268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.599305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.599316 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:37Z","lastTransitionTime":"2025-12-06T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.702740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.702869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.702879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.702894 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.702905 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:37Z","lastTransitionTime":"2025-12-06T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.805474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.805527 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.805539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.805558 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.805576 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:37Z","lastTransitionTime":"2025-12-06T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.908638 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.908685 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.908694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.908712 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.908723 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:37Z","lastTransitionTime":"2025-12-06T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.952549 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.952642 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:37 crc kubenswrapper[4945]: E1206 06:53:37.952735 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:37 crc kubenswrapper[4945]: E1206 06:53:37.952860 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:37 crc kubenswrapper[4945]: I1206 06:53:37.953007 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:37 crc kubenswrapper[4945]: E1206 06:53:37.953101 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.012156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.012208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.012220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.012240 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.012255 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:38Z","lastTransitionTime":"2025-12-06T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.114864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.114900 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.114909 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.114922 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.114933 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:38Z","lastTransitionTime":"2025-12-06T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.210275 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.218593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.218651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.218665 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.218685 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.218696 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:38Z","lastTransitionTime":"2025-12-06T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.321488 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.321536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.321545 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.321565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.321577 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:38Z","lastTransitionTime":"2025-12-06T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.424801 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.424863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.424884 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.424910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.424926 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:38Z","lastTransitionTime":"2025-12-06T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.528165 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.528244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.528266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.528336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.528364 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:38Z","lastTransitionTime":"2025-12-06T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.630828 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.630916 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.630928 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.630950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.630964 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:38Z","lastTransitionTime":"2025-12-06T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.734552 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.734619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.734637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.734669 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.734692 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:38Z","lastTransitionTime":"2025-12-06T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.838431 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.838520 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.838547 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.838582 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.838605 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:38Z","lastTransitionTime":"2025-12-06T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.941765 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.941825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.941843 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.941864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:38 crc kubenswrapper[4945]: I1206 06:53:38.941878 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:38Z","lastTransitionTime":"2025-12-06T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.044316 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.044379 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.044398 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.044426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.044449 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:39Z","lastTransitionTime":"2025-12-06T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.148241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.148378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.148396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.148423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.148439 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:39Z","lastTransitionTime":"2025-12-06T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.216879 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/0.log" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.221501 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958" exitCode=1 Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.221544 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.222229 4945 scope.go:117] "RemoveContainer" containerID="bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.242272 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.251504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.251738 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.251857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.252010 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.252227 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:39Z","lastTransitionTime":"2025-12-06T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.259119 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.270028 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.291475 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.306568 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.320101 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.340770 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.354871 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.354910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.354946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.354965 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.354976 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:39Z","lastTransitionTime":"2025-12-06T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.361154 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.376661 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.389875 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.403456 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.416894 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.432686 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.452519 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.457406 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.457476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.457492 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.457515 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.457530 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:39Z","lastTransitionTime":"2025-12-06T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.482763 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:39Z\\\",\\\"message\\\":\\\"or removal\\\\nI1206 06:53:38.422759 6233 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 06:53:38.422765 6233 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 06:53:38.422850 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 06:53:38.422893 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 06:53:38.423006 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:38.423020 6233 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 06:53:38.423026 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 06:53:38.423034 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 06:53:38.423041 6233 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 06:53:38.423048 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 06:53:38.423054 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:38.424807 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 06:53:38.424884 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 06:53:38.424934 6233 factory.go:656] Stopping watch factory\\\\nI1206 06:53:38.424940 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 06:53:38.424969 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:39Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.560347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.560387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.560396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.560412 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.560424 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:39Z","lastTransitionTime":"2025-12-06T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.663625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.663677 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.663687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.663707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.663720 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:39Z","lastTransitionTime":"2025-12-06T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.766126 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.766174 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.766187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.766204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.766220 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:39Z","lastTransitionTime":"2025-12-06T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.869266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.869343 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.869356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.869378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.869393 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:39Z","lastTransitionTime":"2025-12-06T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.952519 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.952519 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:39 crc kubenswrapper[4945]: E1206 06:53:39.952664 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:39 crc kubenswrapper[4945]: E1206 06:53:39.952725 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.952519 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:39 crc kubenswrapper[4945]: E1206 06:53:39.952826 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.972039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.972078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.972090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.972112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:39 crc kubenswrapper[4945]: I1206 06:53:39.972127 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:39Z","lastTransitionTime":"2025-12-06T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.075135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.075198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.075217 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.075240 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.075258 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:40Z","lastTransitionTime":"2025-12-06T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.178154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.178210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.178221 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.178241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.178252 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:40Z","lastTransitionTime":"2025-12-06T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.228333 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/0.log" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.232088 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88"} Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.232229 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.248124 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.269605 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.281668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.281723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.281734 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.281752 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.281764 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:40Z","lastTransitionTime":"2025-12-06T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.284250 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.298683 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.316156 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.345260 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.372433 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.384320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.384369 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.384381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.384403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.384415 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:40Z","lastTransitionTime":"2025-12-06T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.403596 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.428611 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:39Z\\\",\\\"message\\\":\\\"or removal\\\\nI1206 06:53:38.422759 6233 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 06:53:38.422765 6233 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 06:53:38.422850 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 06:53:38.422893 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 06:53:38.423006 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:38.423020 6233 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 06:53:38.423026 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 06:53:38.423034 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 06:53:38.423041 6233 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 06:53:38.423048 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 06:53:38.423054 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:38.424807 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 06:53:38.424884 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 06:53:38.424934 6233 factory.go:656] Stopping watch factory\\\\nI1206 06:53:38.424940 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 06:53:38.424969 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.444909 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.461611 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.475309 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.487680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.488000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.488069 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.488315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.488419 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:40Z","lastTransitionTime":"2025-12-06T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.492040 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.507252 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.523045 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:40Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.591926 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.591999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.592020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.592048 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.592070 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:40Z","lastTransitionTime":"2025-12-06T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.694914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.694967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.694980 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.694999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.695012 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:40Z","lastTransitionTime":"2025-12-06T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.798638 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.799116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.799194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.799319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.799415 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:40Z","lastTransitionTime":"2025-12-06T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.902563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.902616 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.902631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.902651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:40 crc kubenswrapper[4945]: I1206 06:53:40.902663 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:40Z","lastTransitionTime":"2025-12-06T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.006799 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.006877 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.006895 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.006923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.006942 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:41Z","lastTransitionTime":"2025-12-06T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.109970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.110031 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.110044 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.110064 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.110079 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:41Z","lastTransitionTime":"2025-12-06T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.177031 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz"] Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.177939 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.180752 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.181821 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.196511 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.214142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.214224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.214246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.214302 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.214323 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:41Z","lastTransitionTime":"2025-12-06T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.220579 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.227689 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnsn8\" (UniqueName: \"kubernetes.io/projected/680cc0bb-d74c-42de-8dcc-0690426fdfa5-kube-api-access-pnsn8\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.227768 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/680cc0bb-d74c-42de-8dcc-0690426fdfa5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.227877 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/680cc0bb-d74c-42de-8dcc-0690426fdfa5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.227973 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/680cc0bb-d74c-42de-8dcc-0690426fdfa5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.238857 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/1.log" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.240025 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/0.log" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.243970 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88" exitCode=1 Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.244040 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.244159 4945 scope.go:117] "RemoveContainer" containerID="bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.244776 4945 scope.go:117] "RemoveContainer" containerID="0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88" Dec 06 06:53:41 crc kubenswrapper[4945]: E1206 06:53:41.244963 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.253251 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:39Z\\\",\\\"message\\\":\\\"or removal\\\\nI1206 06:53:38.422759 6233 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 06:53:38.422765 6233 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 06:53:38.422850 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 06:53:38.422893 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 06:53:38.423006 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:38.423020 6233 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 06:53:38.423026 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 06:53:38.423034 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 06:53:38.423041 6233 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 06:53:38.423048 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 06:53:38.423054 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:38.424807 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 06:53:38.424884 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 06:53:38.424934 6233 factory.go:656] Stopping watch factory\\\\nI1206 06:53:38.424940 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 06:53:38.424969 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.268812 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.284625 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.302201 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.312920 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.316841 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.316927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.316951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.316976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.316990 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:41Z","lastTransitionTime":"2025-12-06T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.329037 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/680cc0bb-d74c-42de-8dcc-0690426fdfa5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.329158 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnsn8\" (UniqueName: \"kubernetes.io/projected/680cc0bb-d74c-42de-8dcc-0690426fdfa5-kube-api-access-pnsn8\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.329375 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/680cc0bb-d74c-42de-8dcc-0690426fdfa5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.329419 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/680cc0bb-d74c-42de-8dcc-0690426fdfa5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.330094 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/680cc0bb-d74c-42de-8dcc-0690426fdfa5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.330190 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/680cc0bb-d74c-42de-8dcc-0690426fdfa5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.331150 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.335525 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/680cc0bb-d74c-42de-8dcc-0690426fdfa5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.347887 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.349770 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnsn8\" (UniqueName: \"kubernetes.io/projected/680cc0bb-d74c-42de-8dcc-0690426fdfa5-kube-api-access-pnsn8\") pod \"ovnkube-control-plane-749d76644c-tprpz\" (UID: \"680cc0bb-d74c-42de-8dcc-0690426fdfa5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.363471 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.389060 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.406187 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.422665 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.422982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.423067 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.423090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.423102 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:41Z","lastTransitionTime":"2025-12-06T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.425795 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.439898 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.454083 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.470799 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.490741 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.501765 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.513722 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:39Z\\\",\\\"message\\\":\\\"or removal\\\\nI1206 06:53:38.422759 6233 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 06:53:38.422765 6233 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 06:53:38.422850 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 06:53:38.422893 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 06:53:38.423006 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:38.423020 6233 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 06:53:38.423026 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 06:53:38.423034 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 06:53:38.423041 6233 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 06:53:38.423048 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 06:53:38.423054 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:38.424807 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 06:53:38.424884 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 06:53:38.424934 6233 factory.go:656] Stopping watch factory\\\\nI1206 06:53:38.424940 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 06:53:38.424969 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:40Z\\\",\\\"message\\\":\\\"hed syncing service machine-config-operator on namespace openshift-machine-config-operator for network=default : 1.994718ms\\\\nI1206 06:53:40.113549 6356 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 06:53:40.113586 6356 services_controller.go:356] Processing sync for service openshift-console/console for network=default\\\\nF1206 06:53:40.113598 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: W1206 06:53:41.518188 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod680cc0bb_d74c_42de_8dcc_0690426fdfa5.slice/crio-42328ee7357abbcb32ea887ae7c5188c549a69051fefccda736f29d0620f711c WatchSource:0}: Error finding container 42328ee7357abbcb32ea887ae7c5188c549a69051fefccda736f29d0620f711c: Status 404 returned error can't find the container with id 42328ee7357abbcb32ea887ae7c5188c549a69051fefccda736f29d0620f711c Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.525625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.525662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.525672 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.525689 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.525701 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:41Z","lastTransitionTime":"2025-12-06T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.533295 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.548877 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.564754 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.581213 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.596135 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.611482 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.625412 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.629162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.629201 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.629212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.629230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.629240 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:41Z","lastTransitionTime":"2025-12-06T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.641990 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.658417 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.676378 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.690428 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.704176 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.717234 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.733085 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.733134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.733145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.733167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.733179 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:41Z","lastTransitionTime":"2025-12-06T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.738204 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:41Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.836332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.836426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.836466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.836503 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.836529 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:41Z","lastTransitionTime":"2025-12-06T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.940005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.940060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.940079 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.940107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.940127 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:41Z","lastTransitionTime":"2025-12-06T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.952510 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.952644 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:41 crc kubenswrapper[4945]: E1206 06:53:41.952694 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:41 crc kubenswrapper[4945]: I1206 06:53:41.952532 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:41 crc kubenswrapper[4945]: E1206 06:53:41.952859 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:41 crc kubenswrapper[4945]: E1206 06:53:41.953154 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.043049 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.043099 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.043111 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.043130 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.043144 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:42Z","lastTransitionTime":"2025-12-06T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.145411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.145526 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.145544 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.145566 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.145586 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:42Z","lastTransitionTime":"2025-12-06T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.248167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.248209 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.248219 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.248238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.248271 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:42Z","lastTransitionTime":"2025-12-06T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.250184 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" event={"ID":"680cc0bb-d74c-42de-8dcc-0690426fdfa5","Type":"ContainerStarted","Data":"ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.250275 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" event={"ID":"680cc0bb-d74c-42de-8dcc-0690426fdfa5","Type":"ContainerStarted","Data":"1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.250329 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" event={"ID":"680cc0bb-d74c-42de-8dcc-0690426fdfa5","Type":"ContainerStarted","Data":"42328ee7357abbcb32ea887ae7c5188c549a69051fefccda736f29d0620f711c"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.252545 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/1.log" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.257403 4945 scope.go:117] "RemoveContainer" containerID="0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88" Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.257600 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.268598 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.284453 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.306480 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.325365 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.352127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.352194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.352212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.352235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.352251 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:42Z","lastTransitionTime":"2025-12-06T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.353768 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.368145 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.387193 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.403590 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.419783 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.436510 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.439730 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.439871 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.439955 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:58.439932584 +0000 UTC m=+51.894793628 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.450058 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.454661 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.454907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.455017 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.455167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.455257 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:42Z","lastTransitionTime":"2025-12-06T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.465737 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.480499 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.497149 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.522391 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bff6780d450feac8ed255315e964c66690aa31a47ddc63cfdf804f3f0a839958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:39Z\\\",\\\"message\\\":\\\"or removal\\\\nI1206 06:53:38.422759 6233 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 06:53:38.422765 6233 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 06:53:38.422850 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 06:53:38.422893 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 06:53:38.423006 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:38.423020 6233 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 06:53:38.423026 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 06:53:38.423034 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 06:53:38.423041 6233 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 06:53:38.423048 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 06:53:38.423054 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:38.424807 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 06:53:38.424884 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 06:53:38.424934 6233 factory.go:656] Stopping watch factory\\\\nI1206 06:53:38.424940 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 06:53:38.424969 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:40Z\\\",\\\"message\\\":\\\"hed syncing service machine-config-operator on namespace openshift-machine-config-operator for network=default : 1.994718ms\\\\nI1206 06:53:40.113549 6356 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 06:53:40.113586 6356 services_controller.go:356] Processing sync for service openshift-console/console for network=default\\\\nF1206 06:53:40.113598 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.536985 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.540551 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.540819 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:53:58.540783256 +0000 UTC m=+51.995644340 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.540959 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.541083 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.541139 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.541251 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:58.541226537 +0000 UTC m=+51.996087591 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.541142 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.541337 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.541365 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.541381 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.541442 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.541465 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:58.541443292 +0000 UTC m=+51.996304376 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.541482 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.541512 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.541605 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 06:53:58.541577945 +0000 UTC m=+51.996439079 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.554163 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.558593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.558636 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.558648 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.558669 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.558682 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:42Z","lastTransitionTime":"2025-12-06T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.571785 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.588321 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.614059 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.632722 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.648775 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.661011 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.661065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.661076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.661093 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.661105 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:42Z","lastTransitionTime":"2025-12-06T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.663239 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.674206 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.686862 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.699312 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.714631 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-2xkp9"] Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.714635 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.716936 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.717024 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.734837 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:40Z\\\",\\\"message\\\":\\\"hed syncing service machine-config-operator on namespace openshift-machine-config-operator for network=default : 1.994718ms\\\\nI1206 06:53:40.113549 6356 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 06:53:40.113586 6356 services_controller.go:356] Processing sync for service openshift-console/console for network=default\\\\nF1206 06:53:40.113598 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.742129 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b58mz\" (UniqueName: \"kubernetes.io/projected/d4326ced-ae97-4301-854b-222062d4764e-kube-api-access-b58mz\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.742193 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.747301 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.760476 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.763335 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.763375 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.763388 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.763409 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.763425 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:42Z","lastTransitionTime":"2025-12-06T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.774031 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.784567 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.796900 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.810614 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.824860 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.843291 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b58mz\" (UniqueName: \"kubernetes.io/projected/d4326ced-ae97-4301-854b-222062d4764e-kube-api-access-b58mz\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.843746 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.843944 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:42 crc kubenswrapper[4945]: E1206 06:53:42.844055 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs podName:d4326ced-ae97-4301-854b-222062d4764e nodeName:}" failed. No retries permitted until 2025-12-06 06:53:43.344030979 +0000 UTC m=+36.798892013 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs") pod "network-metrics-daemon-2xkp9" (UID: "d4326ced-ae97-4301-854b-222062d4764e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.845886 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.864262 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.866333 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.866381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.866395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.866413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.866426 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:42Z","lastTransitionTime":"2025-12-06T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.868569 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b58mz\" (UniqueName: \"kubernetes.io/projected/d4326ced-ae97-4301-854b-222062d4764e-kube-api-access-b58mz\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.885768 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.900699 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.913784 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.932030 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.947820 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.966535 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:40Z\\\",\\\"message\\\":\\\"hed syncing service machine-config-operator on namespace openshift-machine-config-operator for network=default : 1.994718ms\\\\nI1206 06:53:40.113549 6356 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 06:53:40.113586 6356 services_controller.go:356] Processing sync for service openshift-console/console for network=default\\\\nF1206 06:53:40.113598 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.968864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.968906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.968920 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.968938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.968951 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:42Z","lastTransitionTime":"2025-12-06T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.980121 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:42 crc kubenswrapper[4945]: I1206 06:53:42.992749 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:42Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.005334 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:43Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.016298 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:43Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.028062 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:43Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.040479 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:43Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.070978 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.071042 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.071055 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.071070 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.071082 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.175154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.175220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.175239 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.175268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.175323 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.254138 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.261593 4945 scope.go:117] "RemoveContainer" containerID="0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88" Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.261791 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.278239 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.278336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.278350 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.278374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.278387 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.349544 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.349833 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.349989 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs podName:d4326ced-ae97-4301-854b-222062d4764e nodeName:}" failed. No retries permitted until 2025-12-06 06:53:44.349953562 +0000 UTC m=+37.804814806 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs") pod "network-metrics-daemon-2xkp9" (UID: "d4326ced-ae97-4301-854b-222062d4764e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.381358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.381414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.381426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.381445 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.381460 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.476376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.476425 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.476435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.476453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.476464 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.490556 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:43Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.495017 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.495103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.495130 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.495170 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.495196 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.518364 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:43Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.523182 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.523229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.523242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.523264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.523295 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.538003 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:43Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.541926 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.542059 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.542169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.542268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.542364 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.555242 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:43Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.559914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.559958 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.559968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.559988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.559999 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.572197 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:43Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.572329 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.574225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.574258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.574269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.574305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.574328 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.676857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.676919 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.676931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.676948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.676960 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.779824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.779880 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.779892 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.779911 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.779928 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.882954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.883023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.883056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.883098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.883123 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.952242 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.952242 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.952346 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.952392 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.952624 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.952705 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.952784 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:43 crc kubenswrapper[4945]: E1206 06:53:43.952865 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.986252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.986317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.986328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.986346 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:43 crc kubenswrapper[4945]: I1206 06:53:43.986358 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:43Z","lastTransitionTime":"2025-12-06T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.089195 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.089260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.089291 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.089312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.089326 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:44Z","lastTransitionTime":"2025-12-06T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.192436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.192518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.192544 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.192571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.192591 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:44Z","lastTransitionTime":"2025-12-06T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.296426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.296495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.296517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.296537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.296557 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:44Z","lastTransitionTime":"2025-12-06T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.359463 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:44 crc kubenswrapper[4945]: E1206 06:53:44.359695 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:44 crc kubenswrapper[4945]: E1206 06:53:44.359805 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs podName:d4326ced-ae97-4301-854b-222062d4764e nodeName:}" failed. No retries permitted until 2025-12-06 06:53:46.359776807 +0000 UTC m=+39.814637861 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs") pod "network-metrics-daemon-2xkp9" (UID: "d4326ced-ae97-4301-854b-222062d4764e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.399543 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.399596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.399605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.399625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.399639 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:44Z","lastTransitionTime":"2025-12-06T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.503056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.503136 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.503156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.503186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.503204 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:44Z","lastTransitionTime":"2025-12-06T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.607064 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.607156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.607190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.607222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.607243 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:44Z","lastTransitionTime":"2025-12-06T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.710373 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.710465 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.710482 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.710506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.710529 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:44Z","lastTransitionTime":"2025-12-06T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.813451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.813504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.813515 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.813529 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.813539 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:44Z","lastTransitionTime":"2025-12-06T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.916204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.916248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.916264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.916307 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:44 crc kubenswrapper[4945]: I1206 06:53:44.916324 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:44Z","lastTransitionTime":"2025-12-06T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.019204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.019269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.019322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.019348 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.019365 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:45Z","lastTransitionTime":"2025-12-06T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.122950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.123025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.123046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.123075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.123093 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:45Z","lastTransitionTime":"2025-12-06T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.226951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.227038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.227057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.227083 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.227101 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:45Z","lastTransitionTime":"2025-12-06T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.330565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.330625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.330638 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.330657 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.330670 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:45Z","lastTransitionTime":"2025-12-06T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.433693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.433736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.433747 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.433766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.433776 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:45Z","lastTransitionTime":"2025-12-06T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.536658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.536697 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.536708 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.536723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.536732 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:45Z","lastTransitionTime":"2025-12-06T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.639895 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.639957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.639975 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.640005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.640020 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:45Z","lastTransitionTime":"2025-12-06T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.742646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.742693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.742704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.742722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.742732 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:45Z","lastTransitionTime":"2025-12-06T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.846495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.846553 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.846571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.846598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.846621 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:45Z","lastTransitionTime":"2025-12-06T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.949452 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.949494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.949504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.949522 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.949532 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:45Z","lastTransitionTime":"2025-12-06T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.953170 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.953198 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.953240 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:45 crc kubenswrapper[4945]: I1206 06:53:45.953233 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:45 crc kubenswrapper[4945]: E1206 06:53:45.953360 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:45 crc kubenswrapper[4945]: E1206 06:53:45.953496 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:45 crc kubenswrapper[4945]: E1206 06:53:45.953582 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:53:45 crc kubenswrapper[4945]: E1206 06:53:45.953746 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.054086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.054159 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.054177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.054195 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.054226 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:46Z","lastTransitionTime":"2025-12-06T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.157344 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.157449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.157468 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.157547 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.157578 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:46Z","lastTransitionTime":"2025-12-06T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.260905 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.260976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.260994 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.261023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.261124 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:46Z","lastTransitionTime":"2025-12-06T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.364214 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.364263 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.364306 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.364328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.364340 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:46Z","lastTransitionTime":"2025-12-06T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:46 crc kubenswrapper[4945]: E1206 06:53:46.384814 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:46 crc kubenswrapper[4945]: E1206 06:53:46.385386 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs podName:d4326ced-ae97-4301-854b-222062d4764e nodeName:}" failed. No retries permitted until 2025-12-06 06:53:50.385325371 +0000 UTC m=+43.840186425 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs") pod "network-metrics-daemon-2xkp9" (UID: "d4326ced-ae97-4301-854b-222062d4764e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.385477 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.467681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.467745 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.467756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.467776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.467792 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:46Z","lastTransitionTime":"2025-12-06T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.570478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.570545 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.570557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.570576 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.570588 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:46Z","lastTransitionTime":"2025-12-06T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.673441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.673503 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.673518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.673539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.673552 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:46Z","lastTransitionTime":"2025-12-06T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.776924 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.777036 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.777057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.777738 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.777827 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:46Z","lastTransitionTime":"2025-12-06T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.881528 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.881585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.881602 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.881628 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.881648 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:46Z","lastTransitionTime":"2025-12-06T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.983238 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:40Z\\\",\\\"message\\\":\\\"hed syncing service machine-config-operator on namespace openshift-machine-config-operator for network=default : 1.994718ms\\\\nI1206 06:53:40.113549 6356 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 06:53:40.113586 6356 services_controller.go:356] Processing sync for service openshift-console/console for network=default\\\\nF1206 06:53:40.113598 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:46Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.984699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.984742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.984757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.984780 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:46 crc kubenswrapper[4945]: I1206 06:53:46.984793 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:46Z","lastTransitionTime":"2025-12-06T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.002268 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:46Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.019156 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.040110 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.057349 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.072911 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.091621 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.091659 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.091671 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.091689 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.091700 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:47Z","lastTransitionTime":"2025-12-06T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.112909 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.138300 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.156155 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.169720 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.183802 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.195191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.195232 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.195243 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.195260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.195271 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:47Z","lastTransitionTime":"2025-12-06T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.195872 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.210404 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.232362 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.250663 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.264254 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.278161 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.297259 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.297320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.297332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.297351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.297363 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:47Z","lastTransitionTime":"2025-12-06T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.400387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.400467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.400486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.401031 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.401092 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:47Z","lastTransitionTime":"2025-12-06T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.504650 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.504752 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.504780 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.504820 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.504844 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:47Z","lastTransitionTime":"2025-12-06T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.608136 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.608190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.608200 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.608216 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.608229 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:47Z","lastTransitionTime":"2025-12-06T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.710976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.711074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.711103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.711128 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.711144 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:47Z","lastTransitionTime":"2025-12-06T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.813985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.814037 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.814046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.814064 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.814075 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:47Z","lastTransitionTime":"2025-12-06T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.918355 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.918401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.918418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.918441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.918456 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:47Z","lastTransitionTime":"2025-12-06T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.953102 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.953147 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.953106 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:47 crc kubenswrapper[4945]: E1206 06:53:47.953308 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:47 crc kubenswrapper[4945]: E1206 06:53:47.953371 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:47 crc kubenswrapper[4945]: E1206 06:53:47.953434 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:53:47 crc kubenswrapper[4945]: I1206 06:53:47.953865 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:47 crc kubenswrapper[4945]: E1206 06:53:47.954094 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.024204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.024300 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.024330 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.024360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.024378 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:48Z","lastTransitionTime":"2025-12-06T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.127749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.127806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.127817 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.127836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.127846 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:48Z","lastTransitionTime":"2025-12-06T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.230447 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.230510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.230524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.230546 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.230559 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:48Z","lastTransitionTime":"2025-12-06T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.333802 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.333858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.333870 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.333889 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.333903 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:48Z","lastTransitionTime":"2025-12-06T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.437048 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.437099 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.437116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.437141 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.437162 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:48Z","lastTransitionTime":"2025-12-06T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.539964 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.540019 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.540031 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.540051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.540063 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:48Z","lastTransitionTime":"2025-12-06T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.642550 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.642615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.642633 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.642656 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.642673 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:48Z","lastTransitionTime":"2025-12-06T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.745585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.745643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.745651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.745670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.745682 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:48Z","lastTransitionTime":"2025-12-06T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.849407 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.849486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.849507 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.849537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.849556 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:48Z","lastTransitionTime":"2025-12-06T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.952628 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.952696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.952716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.952742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:48 crc kubenswrapper[4945]: I1206 06:53:48.952761 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:48Z","lastTransitionTime":"2025-12-06T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.055988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.056098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.056112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.056134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.056150 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:49Z","lastTransitionTime":"2025-12-06T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.159182 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.159267 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.159324 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.159360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.159384 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:49Z","lastTransitionTime":"2025-12-06T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.262803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.262878 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.262899 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.262930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.262953 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:49Z","lastTransitionTime":"2025-12-06T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.365648 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.365709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.365722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.365742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.365757 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:49Z","lastTransitionTime":"2025-12-06T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.468596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.468671 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.468691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.468716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.468734 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:49Z","lastTransitionTime":"2025-12-06T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.571189 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.571264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.571320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.571358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.571383 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:49Z","lastTransitionTime":"2025-12-06T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.674819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.674890 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.674903 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.674922 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.674937 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:49Z","lastTransitionTime":"2025-12-06T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.777746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.777820 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.777838 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.777865 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.777886 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:49Z","lastTransitionTime":"2025-12-06T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.881047 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.881138 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.881158 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.881194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.881212 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:49Z","lastTransitionTime":"2025-12-06T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.953111 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.953159 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.953159 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.953324 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:49 crc kubenswrapper[4945]: E1206 06:53:49.953506 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:53:49 crc kubenswrapper[4945]: E1206 06:53:49.953656 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:49 crc kubenswrapper[4945]: E1206 06:53:49.953912 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:49 crc kubenswrapper[4945]: E1206 06:53:49.954091 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.984808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.984880 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.984898 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.984926 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:49 crc kubenswrapper[4945]: I1206 06:53:49.984946 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:49Z","lastTransitionTime":"2025-12-06T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.087556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.087616 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.087633 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.087668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.087686 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:50Z","lastTransitionTime":"2025-12-06T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.190442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.190501 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.190519 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.190542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.190556 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:50Z","lastTransitionTime":"2025-12-06T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.292231 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.292338 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.292361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.292388 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.292409 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:50Z","lastTransitionTime":"2025-12-06T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.395694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.395793 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.395812 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.395837 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.395860 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:50Z","lastTransitionTime":"2025-12-06T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.441984 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:50 crc kubenswrapper[4945]: E1206 06:53:50.442212 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:50 crc kubenswrapper[4945]: E1206 06:53:50.442353 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs podName:d4326ced-ae97-4301-854b-222062d4764e nodeName:}" failed. No retries permitted until 2025-12-06 06:53:58.442325318 +0000 UTC m=+51.897186372 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs") pod "network-metrics-daemon-2xkp9" (UID: "d4326ced-ae97-4301-854b-222062d4764e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.499112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.499183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.499204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.499230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.499248 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:50Z","lastTransitionTime":"2025-12-06T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.602222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.602309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.602332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.602356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.602376 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:50Z","lastTransitionTime":"2025-12-06T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.705993 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.706072 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.706094 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.706127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.706152 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:50Z","lastTransitionTime":"2025-12-06T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.810168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.810246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.810264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.810348 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.810373 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:50Z","lastTransitionTime":"2025-12-06T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.914171 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.914232 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.914251 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.914312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:50 crc kubenswrapper[4945]: I1206 06:53:50.914334 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:50Z","lastTransitionTime":"2025-12-06T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.017723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.017786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.017795 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.017814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.017826 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:51Z","lastTransitionTime":"2025-12-06T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.121151 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.121215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.121227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.121244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.121256 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:51Z","lastTransitionTime":"2025-12-06T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.224677 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.224750 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.224769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.224796 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.224816 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:51Z","lastTransitionTime":"2025-12-06T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.327756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.327828 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.327844 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.327867 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.327882 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:51Z","lastTransitionTime":"2025-12-06T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.431385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.431476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.431493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.431516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.431533 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:51Z","lastTransitionTime":"2025-12-06T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.534669 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.534729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.534746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.534773 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.534786 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:51Z","lastTransitionTime":"2025-12-06T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.637615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.637701 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.637723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.637866 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.637918 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:51Z","lastTransitionTime":"2025-12-06T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.741492 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.741578 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.741598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.741625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.741642 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:51Z","lastTransitionTime":"2025-12-06T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.844358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.844405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.844414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.844432 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.844441 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:51Z","lastTransitionTime":"2025-12-06T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.946951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.946998 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.947009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.947025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.947036 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:51Z","lastTransitionTime":"2025-12-06T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.952211 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.952308 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.952330 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:51 crc kubenswrapper[4945]: I1206 06:53:51.952487 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:51 crc kubenswrapper[4945]: E1206 06:53:51.952514 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:51 crc kubenswrapper[4945]: E1206 06:53:51.952593 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:51 crc kubenswrapper[4945]: E1206 06:53:51.952718 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:51 crc kubenswrapper[4945]: E1206 06:53:51.952860 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.051434 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.051536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.051564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.051601 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.051626 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:52Z","lastTransitionTime":"2025-12-06T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.154884 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.154966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.154982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.155007 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.155023 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:52Z","lastTransitionTime":"2025-12-06T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.257765 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.257814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.257823 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.257843 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.257854 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:52Z","lastTransitionTime":"2025-12-06T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.360741 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.360782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.360792 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.360806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.360818 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:52Z","lastTransitionTime":"2025-12-06T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.463485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.463567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.463582 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.463599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.463610 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:52Z","lastTransitionTime":"2025-12-06T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.566563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.566622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.566634 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.566655 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.566671 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:52Z","lastTransitionTime":"2025-12-06T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.668866 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.668920 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.668932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.668949 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.668961 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:52Z","lastTransitionTime":"2025-12-06T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.771432 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.771468 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.771476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.771493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.771504 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:52Z","lastTransitionTime":"2025-12-06T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.874695 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.874732 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.874741 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.874756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.874767 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:52Z","lastTransitionTime":"2025-12-06T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.977092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.977145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.977158 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.977180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:52 crc kubenswrapper[4945]: I1206 06:53:52.977194 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:52Z","lastTransitionTime":"2025-12-06T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.080419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.080468 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.080479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.080496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.080507 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.182412 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.182453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.182463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.182482 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.182493 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.284686 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.284728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.284736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.284751 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.284761 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.388614 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.388682 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.388702 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.388734 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.388758 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.491766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.491872 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.491890 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.491914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.491931 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.595264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.596489 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.596670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.596815 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.596938 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.700427 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.700528 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.700556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.700594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.700618 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.716609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.716681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.716695 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.716716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.716735 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: E1206 06:53:53.737984 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:53Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.743882 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.743936 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.743953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.743977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.743996 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: E1206 06:53:53.766174 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:53Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.772313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.772351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.772360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.772377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.772388 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: E1206 06:53:53.790185 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:53Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.795412 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.795484 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.795504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.795531 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.795552 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: E1206 06:53:53.822922 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:53Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.829265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.829376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.829399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.829427 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.829446 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: E1206 06:53:53.846105 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:53Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:53 crc kubenswrapper[4945]: E1206 06:53:53.846308 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.848619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.848678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.848696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.848722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.848785 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.951948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.952019 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.952037 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.952065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.952082 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:53Z","lastTransitionTime":"2025-12-06T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.952451 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.952509 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.952471 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:53 crc kubenswrapper[4945]: I1206 06:53:53.952556 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:53 crc kubenswrapper[4945]: E1206 06:53:53.952611 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:53 crc kubenswrapper[4945]: E1206 06:53:53.952784 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:53:53 crc kubenswrapper[4945]: E1206 06:53:53.952909 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:53 crc kubenswrapper[4945]: E1206 06:53:53.953109 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.055981 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.056053 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.056074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.056101 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.056121 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:54Z","lastTransitionTime":"2025-12-06T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.160460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.160532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.160551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.160577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.160605 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:54Z","lastTransitionTime":"2025-12-06T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.264603 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.264675 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.264694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.264721 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.264741 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:54Z","lastTransitionTime":"2025-12-06T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.366870 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.366902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.366910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.366924 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.366935 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:54Z","lastTransitionTime":"2025-12-06T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.469897 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.469999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.470024 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.470058 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.470077 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:54Z","lastTransitionTime":"2025-12-06T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.573506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.573594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.573620 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.573644 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.573656 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:54Z","lastTransitionTime":"2025-12-06T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.677335 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.677407 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.677433 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.677466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.677490 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:54Z","lastTransitionTime":"2025-12-06T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.780581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.780631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.780643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.780660 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.780673 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:54Z","lastTransitionTime":"2025-12-06T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.883756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.883844 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.883868 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.883895 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.883918 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:54Z","lastTransitionTime":"2025-12-06T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.988358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.988432 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.988450 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.988477 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:54 crc kubenswrapper[4945]: I1206 06:53:54.988496 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:54Z","lastTransitionTime":"2025-12-06T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.092186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.092259 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.092312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.092341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.092360 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:55Z","lastTransitionTime":"2025-12-06T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.195961 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.196018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.196036 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.196061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.196080 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:55Z","lastTransitionTime":"2025-12-06T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.299444 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.299525 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.299547 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.299579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.299598 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:55Z","lastTransitionTime":"2025-12-06T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.403892 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.403972 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.403995 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.404029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.404053 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:55Z","lastTransitionTime":"2025-12-06T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.507733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.507829 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.507858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.507892 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.507914 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:55Z","lastTransitionTime":"2025-12-06T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.610944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.610980 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.610989 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.611004 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.611016 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:55Z","lastTransitionTime":"2025-12-06T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.714154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.714215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.714227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.714248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.714261 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:55Z","lastTransitionTime":"2025-12-06T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.818161 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.818245 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.818261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.818325 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.818342 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:55Z","lastTransitionTime":"2025-12-06T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.921751 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.921823 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.921839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.921862 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.921877 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:55Z","lastTransitionTime":"2025-12-06T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.953428 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.953500 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.953506 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:55 crc kubenswrapper[4945]: I1206 06:53:55.953701 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:55 crc kubenswrapper[4945]: E1206 06:53:55.953923 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:55 crc kubenswrapper[4945]: E1206 06:53:55.954093 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:55 crc kubenswrapper[4945]: E1206 06:53:55.954317 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:53:55 crc kubenswrapper[4945]: E1206 06:53:55.954544 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.025047 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.025116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.025133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.025160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.025179 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:56Z","lastTransitionTime":"2025-12-06T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.128327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.128380 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.128390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.128406 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.128422 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:56Z","lastTransitionTime":"2025-12-06T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.231329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.231378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.231386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.231402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.231412 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:56Z","lastTransitionTime":"2025-12-06T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.334594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.334698 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.334728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.334767 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.334789 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:56Z","lastTransitionTime":"2025-12-06T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.437727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.437787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.437804 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.437826 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.437841 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:56Z","lastTransitionTime":"2025-12-06T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.540198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.540233 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.540243 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.540258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.540267 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:56Z","lastTransitionTime":"2025-12-06T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.644099 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.644147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.644158 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.644177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.644190 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:56Z","lastTransitionTime":"2025-12-06T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.746688 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.746746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.746762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.746783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.746799 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:56Z","lastTransitionTime":"2025-12-06T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.850125 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.850173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.850188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.850226 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.850252 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:56Z","lastTransitionTime":"2025-12-06T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.953563 4945 scope.go:117] "RemoveContainer" containerID="0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.954655 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.954757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.954779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.954803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.954865 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:56Z","lastTransitionTime":"2025-12-06T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.975380 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:56Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:56 crc kubenswrapper[4945]: I1206 06:53:56.996522 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:56Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.012480 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.025123 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.040480 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.054831 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.058006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.058044 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.058056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.058075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.058088 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:57Z","lastTransitionTime":"2025-12-06T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.070381 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.092985 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.110041 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.127095 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.139917 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.152921 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.161088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.161151 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.161170 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.161191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.161204 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:57Z","lastTransitionTime":"2025-12-06T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.165801 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.181667 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.198506 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.230514 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:40Z\\\",\\\"message\\\":\\\"hed syncing service machine-config-operator on namespace openshift-machine-config-operator for network=default : 1.994718ms\\\\nI1206 06:53:40.113549 6356 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 06:53:40.113586 6356 services_controller.go:356] Processing sync for service openshift-console/console for network=default\\\\nF1206 06:53:40.113598 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.244774 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.264420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.264459 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.264471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.264486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.264499 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:57Z","lastTransitionTime":"2025-12-06T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.314419 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/1.log" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.317523 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97"} Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.318083 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.342979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.361560 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.367435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.367486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.367500 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.367522 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.367537 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:57Z","lastTransitionTime":"2025-12-06T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.383487 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.397085 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.414220 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.429952 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.444568 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.460544 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.470364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.470407 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.470423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.470444 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.470458 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:57Z","lastTransitionTime":"2025-12-06T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.480447 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:40Z\\\",\\\"message\\\":\\\"hed syncing service machine-config-operator on namespace openshift-machine-config-operator for network=default : 1.994718ms\\\\nI1206 06:53:40.113549 6356 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 06:53:40.113586 6356 services_controller.go:356] Processing sync for service openshift-console/console for network=default\\\\nF1206 06:53:40.113598 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.493599 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.507338 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.522821 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.534546 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.546392 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.562391 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.572816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.572863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.572874 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.572894 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.572908 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:57Z","lastTransitionTime":"2025-12-06T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.577506 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.596572 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:57Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.676251 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.676322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.676335 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.676354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.676368 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:57Z","lastTransitionTime":"2025-12-06T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.779731 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.779779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.779796 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.779819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.779831 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:57Z","lastTransitionTime":"2025-12-06T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.883141 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.883184 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.883198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.883217 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.883227 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:57Z","lastTransitionTime":"2025-12-06T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.952858 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.952915 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.952877 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:57 crc kubenswrapper[4945]: E1206 06:53:57.953066 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.953136 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:57 crc kubenswrapper[4945]: E1206 06:53:57.953224 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:57 crc kubenswrapper[4945]: E1206 06:53:57.953259 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:57 crc kubenswrapper[4945]: E1206 06:53:57.953373 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.986389 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.986467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.986494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.986526 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:57 crc kubenswrapper[4945]: I1206 06:53:57.986550 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:57Z","lastTransitionTime":"2025-12-06T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.089842 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.089901 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.089918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.089946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.089966 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:58Z","lastTransitionTime":"2025-12-06T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.197419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.197509 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.197542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.197578 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.197603 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:58Z","lastTransitionTime":"2025-12-06T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.301358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.301912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.301989 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.302073 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.302135 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:58Z","lastTransitionTime":"2025-12-06T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.324631 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/2.log" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.326467 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/1.log" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.331249 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97" exitCode=1 Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.331351 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.331408 4945 scope.go:117] "RemoveContainer" containerID="0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.332517 4945 scope.go:117] "RemoveContainer" containerID="7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97" Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.332830 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.353790 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.374542 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.400170 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:40Z\\\",\\\"message\\\":\\\"hed syncing service machine-config-operator on namespace openshift-machine-config-operator for network=default : 1.994718ms\\\\nI1206 06:53:40.113549 6356 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 06:53:40.113586 6356 services_controller.go:356] Processing sync for service openshift-console/console for network=default\\\\nF1206 06:53:40.113598 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"message\\\":\\\":57.859558 6576 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:57.859756 6576 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.859903 6576 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:57.859916 6576 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860016 6576 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.859968 6576 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 06:53:57.859992 6576 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.860260 6576 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860716 6576 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.405650 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.405718 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.405738 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.405765 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.405784 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:58Z","lastTransitionTime":"2025-12-06T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.418637 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.434989 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.452783 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.464815 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.476787 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.496197 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.508462 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.508504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.508512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.508530 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.508540 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:58Z","lastTransitionTime":"2025-12-06T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.516370 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.532276 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.534805 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.534850 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.535019 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.535024 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.535087 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs podName:d4326ced-ae97-4301-854b-222062d4764e nodeName:}" failed. No retries permitted until 2025-12-06 06:54:14.535069793 +0000 UTC m=+67.989930847 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs") pod "network-metrics-daemon-2xkp9" (UID: "d4326ced-ae97-4301-854b-222062d4764e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.535147 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:54:30.535116074 +0000 UTC m=+83.989977138 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.556646 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.572884 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.590779 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.608244 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.611462 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.611516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.611532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.611554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.611571 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:58Z","lastTransitionTime":"2025-12-06T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.623572 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.635802 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.635958 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.635989 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.636014 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636085 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:54:30.636037098 +0000 UTC m=+84.090898162 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636152 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636176 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636191 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636253 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 06:54:30.636233773 +0000 UTC m=+84.091094817 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636310 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636395 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636421 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:54:30.636390037 +0000 UTC m=+84.091251121 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636435 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636459 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:58 crc kubenswrapper[4945]: E1206 06:53:58.636526 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 06:54:30.63651241 +0000 UTC m=+84.091373464 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.640317 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.715354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.715429 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.715453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.715485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.715510 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:58Z","lastTransitionTime":"2025-12-06T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.819300 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.819361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.819375 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.819396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.819409 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:58Z","lastTransitionTime":"2025-12-06T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.835251 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.852078 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.856586 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.875319 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.891495 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.923002 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.923073 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.923093 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.923121 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.923142 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:58Z","lastTransitionTime":"2025-12-06T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.926788 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.952120 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:58 crc kubenswrapper[4945]: I1206 06:53:58.980768 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:58Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.006160 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.026136 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.026517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.026597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.026672 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.026801 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:59Z","lastTransitionTime":"2025-12-06T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.034591 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da11faa1e13f8930e09054adbce5f5db697888d9f838553a484e676d8a99b88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:40Z\\\",\\\"message\\\":\\\"hed syncing service machine-config-operator on namespace openshift-machine-config-operator for network=default : 1.994718ms\\\\nI1206 06:53:40.113549 6356 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 06:53:40.113586 6356 services_controller.go:356] Processing sync for service openshift-console/console for network=default\\\\nF1206 06:53:40.113598 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"message\\\":\\\":57.859558 6576 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:57.859756 6576 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.859903 6576 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:57.859916 6576 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860016 6576 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.859968 6576 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 06:53:57.859992 6576 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.860260 6576 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860716 6576 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.051678 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.065870 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.080886 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.102358 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.129833 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.130809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.130945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.130968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.130993 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.131010 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:59Z","lastTransitionTime":"2025-12-06T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.149006 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.172496 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.202044 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.224500 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.234196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.234367 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.234397 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.234431 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.234450 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:59Z","lastTransitionTime":"2025-12-06T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.337140 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.337961 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.337972 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.337988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.337997 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:59Z","lastTransitionTime":"2025-12-06T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.338294 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/2.log" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.342914 4945 scope.go:117] "RemoveContainer" containerID="7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97" Dec 06 06:53:59 crc kubenswrapper[4945]: E1206 06:53:59.343129 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.358424 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.372237 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.388014 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.412909 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.432920 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.441171 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.441233 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.441244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.441263 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.441304 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:59Z","lastTransitionTime":"2025-12-06T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.449697 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.467437 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.487915 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"message\\\":\\\":57.859558 6576 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:57.859756 6576 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.859903 6576 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:57.859916 6576 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860016 6576 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.859968 6576 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 06:53:57.859992 6576 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.860260 6576 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860716 6576 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.506174 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.522759 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.537415 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4815516-b4c5-469b-8709-75608f08ed9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.543354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.543410 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.543426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.543446 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.543458 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:59Z","lastTransitionTime":"2025-12-06T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.550112 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.563766 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.578469 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.589331 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.604435 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.621917 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.635110 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:53:59Z is after 2025-08-24T17:21:41Z" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.646235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.646269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.646512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.646568 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.646585 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:59Z","lastTransitionTime":"2025-12-06T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.749225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.749299 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.749311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.749330 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.749340 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:59Z","lastTransitionTime":"2025-12-06T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.852057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.852095 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.852103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.852119 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.852129 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:59Z","lastTransitionTime":"2025-12-06T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.952529 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.952577 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.952573 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.952594 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:53:59 crc kubenswrapper[4945]: E1206 06:53:59.952690 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:53:59 crc kubenswrapper[4945]: E1206 06:53:59.952796 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:53:59 crc kubenswrapper[4945]: E1206 06:53:59.952907 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:53:59 crc kubenswrapper[4945]: E1206 06:53:59.953004 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.954449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.954493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.954509 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.954531 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:53:59 crc kubenswrapper[4945]: I1206 06:53:59.954548 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:53:59Z","lastTransitionTime":"2025-12-06T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.057434 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.057494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.057511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.057537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.057555 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:00Z","lastTransitionTime":"2025-12-06T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.160479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.160554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.160577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.160608 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.160631 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:00Z","lastTransitionTime":"2025-12-06T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.263991 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.264061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.264084 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.264117 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.264141 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:00Z","lastTransitionTime":"2025-12-06T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.367670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.367742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.367760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.367789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.367813 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:00Z","lastTransitionTime":"2025-12-06T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.470951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.471004 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.471021 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.471046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.471063 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:00Z","lastTransitionTime":"2025-12-06T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.575114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.575188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.575206 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.575237 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.575259 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:00Z","lastTransitionTime":"2025-12-06T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.678962 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.679045 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.679063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.679093 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.679111 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:00Z","lastTransitionTime":"2025-12-06T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.783186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.783258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.783311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.783340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.783359 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:00Z","lastTransitionTime":"2025-12-06T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.886229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.886325 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.886342 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.886364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.886377 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:00Z","lastTransitionTime":"2025-12-06T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.989556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.989610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.989622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.989641 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:00 crc kubenswrapper[4945]: I1206 06:54:00.989655 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:00Z","lastTransitionTime":"2025-12-06T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.092310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.092367 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.092376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.092395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.092405 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:01Z","lastTransitionTime":"2025-12-06T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.195667 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.195754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.195779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.195816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.195843 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:01Z","lastTransitionTime":"2025-12-06T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.299625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.299683 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.299700 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.299726 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.299748 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:01Z","lastTransitionTime":"2025-12-06T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.403063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.403126 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.403140 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.403160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.403172 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:01Z","lastTransitionTime":"2025-12-06T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.505627 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.505970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.506042 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.506161 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.506267 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:01Z","lastTransitionTime":"2025-12-06T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.610128 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.610652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.610832 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.610990 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.611137 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:01Z","lastTransitionTime":"2025-12-06T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.714531 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.714575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.714587 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.714610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.714626 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:01Z","lastTransitionTime":"2025-12-06T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.818136 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.818224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.818248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.818326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.818356 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:01Z","lastTransitionTime":"2025-12-06T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.922084 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.922130 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.922138 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.922155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.922165 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:01Z","lastTransitionTime":"2025-12-06T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.952300 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.952372 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.952475 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:01 crc kubenswrapper[4945]: E1206 06:54:01.952591 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:01 crc kubenswrapper[4945]: I1206 06:54:01.952618 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:01 crc kubenswrapper[4945]: E1206 06:54:01.952811 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:01 crc kubenswrapper[4945]: E1206 06:54:01.952953 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:01 crc kubenswrapper[4945]: E1206 06:54:01.953093 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.025460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.025552 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.025575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.025605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.025625 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:02Z","lastTransitionTime":"2025-12-06T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.129060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.129149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.129173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.129203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.129224 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:02Z","lastTransitionTime":"2025-12-06T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.232107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.232207 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.232242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.232320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.232353 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:02Z","lastTransitionTime":"2025-12-06T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.334843 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.334899 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.334911 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.334931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.334947 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:02Z","lastTransitionTime":"2025-12-06T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.438102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.438175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.438195 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.438219 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.438237 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:02Z","lastTransitionTime":"2025-12-06T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.542137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.542214 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.542226 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.542247 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.542268 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:02Z","lastTransitionTime":"2025-12-06T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.645892 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.645940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.645952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.645973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.645986 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:02Z","lastTransitionTime":"2025-12-06T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.750567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.750642 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.750662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.750694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.750717 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:02Z","lastTransitionTime":"2025-12-06T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.854084 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.854154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.854173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.854202 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.854223 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:02Z","lastTransitionTime":"2025-12-06T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.957756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.958407 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.958446 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.958476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:02 crc kubenswrapper[4945]: I1206 06:54:02.958496 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:02Z","lastTransitionTime":"2025-12-06T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.061854 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.061904 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.061936 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.061956 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.061966 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:03Z","lastTransitionTime":"2025-12-06T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.165480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.165561 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.165575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.165601 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.165615 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:03Z","lastTransitionTime":"2025-12-06T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.269332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.269405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.269417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.269436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.269453 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:03Z","lastTransitionTime":"2025-12-06T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.372512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.373076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.373346 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.373585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.373792 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:03Z","lastTransitionTime":"2025-12-06T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.477579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.477687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.477710 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.477739 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.477760 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:03Z","lastTransitionTime":"2025-12-06T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.581224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.581311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.581329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.581354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.581384 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:03Z","lastTransitionTime":"2025-12-06T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.685424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.685502 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.685525 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.685561 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.685641 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:03Z","lastTransitionTime":"2025-12-06T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.788675 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.788744 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.788762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.788786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.788806 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:03Z","lastTransitionTime":"2025-12-06T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.891134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.891183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.891197 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.891214 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.891226 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:03Z","lastTransitionTime":"2025-12-06T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.952847 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.952924 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.953033 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:03 crc kubenswrapper[4945]: E1206 06:54:03.953175 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.953373 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:03 crc kubenswrapper[4945]: E1206 06:54:03.953460 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:03 crc kubenswrapper[4945]: E1206 06:54:03.953608 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:03 crc kubenswrapper[4945]: E1206 06:54:03.953758 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.994448 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.994509 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.994530 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.994570 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:03 crc kubenswrapper[4945]: I1206 06:54:03.994595 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:03Z","lastTransitionTime":"2025-12-06T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.017178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.017218 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.017228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.017242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.017254 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: E1206 06:54:04.031957 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:04Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.039819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.039926 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.039948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.040015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.040040 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: E1206 06:54:04.055308 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:04Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.060296 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.060376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.060395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.060415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.060430 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: E1206 06:54:04.074944 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:04Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.080508 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.080564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.080576 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.080596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.080610 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: E1206 06:54:04.095843 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:04Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.100637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.100682 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.100698 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.100716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.100732 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: E1206 06:54:04.117549 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:04Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:04 crc kubenswrapper[4945]: E1206 06:54:04.117715 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.120218 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.120273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.120323 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.120348 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.120367 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.223733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.223809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.223830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.223859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.223879 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.327518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.327701 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.327735 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.327762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.327779 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.430774 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.430806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.430815 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.430829 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.430838 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.534760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.534855 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.534878 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.534911 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.534936 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.637758 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.637802 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.637811 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.637827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.637838 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.741181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.741265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.741320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.741347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.741366 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.845594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.845704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.845720 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.845749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.845767 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.949322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.949386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.949401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.949421 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:04 crc kubenswrapper[4945]: I1206 06:54:04.949436 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:04Z","lastTransitionTime":"2025-12-06T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.053071 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.053144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.053162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.053190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.053211 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:05Z","lastTransitionTime":"2025-12-06T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.157177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.157253 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.157266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.157321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.157335 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:05Z","lastTransitionTime":"2025-12-06T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.260879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.260930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.260939 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.260959 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.260971 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:05Z","lastTransitionTime":"2025-12-06T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.363501 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.363554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.363567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.363585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.363598 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:05Z","lastTransitionTime":"2025-12-06T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.469450 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.469499 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.469512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.469531 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.469543 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:05Z","lastTransitionTime":"2025-12-06T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.572776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.572816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.572825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.572841 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.572852 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:05Z","lastTransitionTime":"2025-12-06T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.676413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.676487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.676505 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.676529 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.676545 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:05Z","lastTransitionTime":"2025-12-06T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.780863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.780925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.780937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.780955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.780967 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:05Z","lastTransitionTime":"2025-12-06T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.884519 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.884592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.884609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.884637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.884655 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:05Z","lastTransitionTime":"2025-12-06T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.952588 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.952643 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.952835 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:05 crc kubenswrapper[4945]: E1206 06:54:05.952825 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:05 crc kubenswrapper[4945]: E1206 06:54:05.953035 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.953092 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:05 crc kubenswrapper[4945]: E1206 06:54:05.953135 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:05 crc kubenswrapper[4945]: E1206 06:54:05.953263 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.987840 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.987908 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.987939 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.987971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:05 crc kubenswrapper[4945]: I1206 06:54:05.987996 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:05Z","lastTransitionTime":"2025-12-06T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.091579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.091662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.091681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.091707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.091725 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:06Z","lastTransitionTime":"2025-12-06T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.194816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.194876 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.194892 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.194916 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.194933 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:06Z","lastTransitionTime":"2025-12-06T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.298244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.298390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.298487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.298526 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.298556 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:06Z","lastTransitionTime":"2025-12-06T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.401571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.401670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.401696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.401733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.401762 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:06Z","lastTransitionTime":"2025-12-06T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.503862 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.503918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.503935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.503960 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.503978 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:06Z","lastTransitionTime":"2025-12-06T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.606647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.606694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.606703 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.606717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.606727 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:06Z","lastTransitionTime":"2025-12-06T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.710026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.710086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.710097 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.710116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.710127 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:06Z","lastTransitionTime":"2025-12-06T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.813663 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.813727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.813740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.813774 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.813791 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:06Z","lastTransitionTime":"2025-12-06T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.916578 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.916639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.916651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.916670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.916683 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:06Z","lastTransitionTime":"2025-12-06T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.975862 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:06Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:06 crc kubenswrapper[4945]: I1206 06:54:06.990291 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4815516-b4c5-469b-8709-75608f08ed9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:06Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.014772 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.020887 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.020943 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.020963 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.020989 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.021010 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:07Z","lastTransitionTime":"2025-12-06T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.040495 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"message\\\":\\\":57.859558 6576 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:57.859756 6576 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.859903 6576 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:57.859916 6576 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860016 6576 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.859968 6576 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 06:53:57.859992 6576 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.860260 6576 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860716 6576 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.058121 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.072817 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.086314 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.097400 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.111911 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.123506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.123541 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.123554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.123572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.123587 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:07Z","lastTransitionTime":"2025-12-06T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.130249 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.145255 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.158169 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.183591 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.201728 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.221731 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.225914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.225950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.225961 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.225981 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.225996 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:07Z","lastTransitionTime":"2025-12-06T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.238665 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.250779 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.266668 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:07Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.331658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.331719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.331751 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.331779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.331796 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:07Z","lastTransitionTime":"2025-12-06T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.435302 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.435352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.435364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.435386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.435399 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:07Z","lastTransitionTime":"2025-12-06T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.538514 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.538561 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.538584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.538603 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.538614 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:07Z","lastTransitionTime":"2025-12-06T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.642324 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.642362 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.642370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.642386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.642395 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:07Z","lastTransitionTime":"2025-12-06T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.745437 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.745489 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.745499 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.745518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.745533 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:07Z","lastTransitionTime":"2025-12-06T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.850946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.851019 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.851039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.851071 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.851097 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:07Z","lastTransitionTime":"2025-12-06T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.952704 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.952787 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.953069 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:07 crc kubenswrapper[4945]: E1206 06:54:07.953195 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:07 crc kubenswrapper[4945]: E1206 06:54:07.953380 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:07 crc kubenswrapper[4945]: E1206 06:54:07.953568 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.953784 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:07 crc kubenswrapper[4945]: E1206 06:54:07.954159 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.955025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.955302 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.955515 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.955748 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:07 crc kubenswrapper[4945]: I1206 06:54:07.956004 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:07Z","lastTransitionTime":"2025-12-06T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.059441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.059496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.059510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.059532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.059544 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:08Z","lastTransitionTime":"2025-12-06T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.163072 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.163130 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.163144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.163169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.163184 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:08Z","lastTransitionTime":"2025-12-06T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.266646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.266697 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.266711 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.266732 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.266746 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:08Z","lastTransitionTime":"2025-12-06T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.371658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.371745 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.371767 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.371799 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.371822 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:08Z","lastTransitionTime":"2025-12-06T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.476017 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.476088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.476103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.476127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.476145 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:08Z","lastTransitionTime":"2025-12-06T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.579668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.579733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.579746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.579767 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.579780 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:08Z","lastTransitionTime":"2025-12-06T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.682809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.683115 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.683196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.683304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.683375 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:08Z","lastTransitionTime":"2025-12-06T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.787161 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.787213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.787225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.787247 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.787269 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:08Z","lastTransitionTime":"2025-12-06T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.890599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.890685 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.890707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.890742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.890766 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:08Z","lastTransitionTime":"2025-12-06T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.995018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.995103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.995126 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.995160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:08 crc kubenswrapper[4945]: I1206 06:54:08.995188 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:08Z","lastTransitionTime":"2025-12-06T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.098815 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.098867 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.098879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.098897 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.098908 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:09Z","lastTransitionTime":"2025-12-06T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.201835 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.201922 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.201975 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.202005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.202020 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:09Z","lastTransitionTime":"2025-12-06T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.305827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.305877 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.305888 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.305910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.305922 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:09Z","lastTransitionTime":"2025-12-06T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.408351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.408393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.408405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.408423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.408439 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:09Z","lastTransitionTime":"2025-12-06T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.521839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.521914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.521947 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.521979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.522002 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:09Z","lastTransitionTime":"2025-12-06T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.625483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.625542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.625560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.625586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.625605 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:09Z","lastTransitionTime":"2025-12-06T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.728904 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.728956 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.728970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.728991 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.729005 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:09Z","lastTransitionTime":"2025-12-06T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.840666 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.840712 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.840723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.840740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.840751 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:09Z","lastTransitionTime":"2025-12-06T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.943858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.943908 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.943918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.943935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.943947 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:09Z","lastTransitionTime":"2025-12-06T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.952982 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.953134 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.953189 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:09 crc kubenswrapper[4945]: I1206 06:54:09.953244 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:09 crc kubenswrapper[4945]: E1206 06:54:09.953373 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:09 crc kubenswrapper[4945]: E1206 06:54:09.953488 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:09 crc kubenswrapper[4945]: E1206 06:54:09.953589 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:09 crc kubenswrapper[4945]: E1206 06:54:09.953711 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.047353 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.047396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.047404 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.047419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.047431 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:10Z","lastTransitionTime":"2025-12-06T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.150300 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.150343 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.150354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.150372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.150385 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:10Z","lastTransitionTime":"2025-12-06T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.253523 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.253572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.253586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.253602 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.253613 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:10Z","lastTransitionTime":"2025-12-06T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.356628 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.356668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.356677 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.356692 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.356702 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:10Z","lastTransitionTime":"2025-12-06T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.459298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.459354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.459364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.459382 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.459393 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:10Z","lastTransitionTime":"2025-12-06T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.563051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.563131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.563156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.563184 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.563204 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:10Z","lastTransitionTime":"2025-12-06T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.666589 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.667145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.667159 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.667180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.667195 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:10Z","lastTransitionTime":"2025-12-06T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.770646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.770740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.770769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.770805 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.770828 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:10Z","lastTransitionTime":"2025-12-06T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.873960 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.874032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.874050 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.874076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.874096 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:10Z","lastTransitionTime":"2025-12-06T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.977105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.977150 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.977163 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.977181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:10 crc kubenswrapper[4945]: I1206 06:54:10.977196 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:10Z","lastTransitionTime":"2025-12-06T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.079891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.079933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.079948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.079967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.079984 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:11Z","lastTransitionTime":"2025-12-06T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.182450 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.182500 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.182513 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.182530 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.182544 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:11Z","lastTransitionTime":"2025-12-06T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.285592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.285635 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.285645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.285661 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.285673 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:11Z","lastTransitionTime":"2025-12-06T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.388419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.388476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.388491 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.388510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.388524 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:11Z","lastTransitionTime":"2025-12-06T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.491542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.492117 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.492147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.492177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.492197 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:11Z","lastTransitionTime":"2025-12-06T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.595670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.595754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.595782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.595850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.595877 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:11Z","lastTransitionTime":"2025-12-06T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.698629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.698678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.698692 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.698710 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.698721 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:11Z","lastTransitionTime":"2025-12-06T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.801874 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.801925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.801935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.801955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.801967 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:11Z","lastTransitionTime":"2025-12-06T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.904359 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.904401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.904415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.904434 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.904448 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:11Z","lastTransitionTime":"2025-12-06T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.952457 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.952531 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.952493 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:11 crc kubenswrapper[4945]: I1206 06:54:11.952699 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:11 crc kubenswrapper[4945]: E1206 06:54:11.952677 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:11 crc kubenswrapper[4945]: E1206 06:54:11.952823 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:11 crc kubenswrapper[4945]: E1206 06:54:11.952918 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:11 crc kubenswrapper[4945]: E1206 06:54:11.953064 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.007070 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.007115 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.007125 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.007144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.007155 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:12Z","lastTransitionTime":"2025-12-06T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.109994 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.110057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.110073 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.110096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.110110 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:12Z","lastTransitionTime":"2025-12-06T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.213361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.213804 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.213953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.214090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.214228 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:12Z","lastTransitionTime":"2025-12-06T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.316607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.317377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.317486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.317580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.317661 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:12Z","lastTransitionTime":"2025-12-06T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.421111 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.421196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.421215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.421246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.421273 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:12Z","lastTransitionTime":"2025-12-06T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.524359 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.524424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.524439 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.524463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.524477 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:12Z","lastTransitionTime":"2025-12-06T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.627211 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.627307 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.627320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.627341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.627354 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:12Z","lastTransitionTime":"2025-12-06T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.730121 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.730173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.730188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.730213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.730232 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:12Z","lastTransitionTime":"2025-12-06T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.833839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.833887 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.833900 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.833918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.833933 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:12Z","lastTransitionTime":"2025-12-06T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.937367 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.937417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.937426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.937441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.937454 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:12Z","lastTransitionTime":"2025-12-06T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:12 crc kubenswrapper[4945]: I1206 06:54:12.954898 4945 scope.go:117] "RemoveContainer" containerID="7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97" Dec 06 06:54:12 crc kubenswrapper[4945]: E1206 06:54:12.955616 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.040654 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.040709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.040722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.040742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.040754 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:13Z","lastTransitionTime":"2025-12-06T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.143633 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.143709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.143729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.143749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.143762 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:13Z","lastTransitionTime":"2025-12-06T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.246657 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.246728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.246747 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.246773 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.246793 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:13Z","lastTransitionTime":"2025-12-06T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.349732 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.349805 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.349822 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.349850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.349867 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:13Z","lastTransitionTime":"2025-12-06T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.453068 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.453116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.453131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.453149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.453159 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:13Z","lastTransitionTime":"2025-12-06T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.555321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.555383 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.555399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.555429 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.555443 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:13Z","lastTransitionTime":"2025-12-06T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.658144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.658178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.658190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.658211 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.658230 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:13Z","lastTransitionTime":"2025-12-06T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.761111 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.761167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.761179 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.761200 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.761212 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:13Z","lastTransitionTime":"2025-12-06T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.864564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.864639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.864649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.864668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.864681 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:13Z","lastTransitionTime":"2025-12-06T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.952715 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.952767 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.952863 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.952901 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:13 crc kubenswrapper[4945]: E1206 06:54:13.953040 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:13 crc kubenswrapper[4945]: E1206 06:54:13.953313 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:13 crc kubenswrapper[4945]: E1206 06:54:13.953567 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:13 crc kubenswrapper[4945]: E1206 06:54:13.953822 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.967468 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.967727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.967840 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.967860 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.967886 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:13 crc kubenswrapper[4945]: I1206 06:54:13.967905 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:13Z","lastTransitionTime":"2025-12-06T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.071041 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.071083 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.071095 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.071112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.071123 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.137196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.137251 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.137265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.137310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.137328 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: E1206 06:54:14.153351 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:14Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.156891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.156933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.156944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.156966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.156979 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: E1206 06:54:14.168850 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:14Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.174014 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.174086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.174105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.174137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.174170 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: E1206 06:54:14.188505 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:14Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.193198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.193270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.193310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.193334 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.193348 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: E1206 06:54:14.205988 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:14Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.210089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.210134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.210149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.210169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.210183 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: E1206 06:54:14.225305 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:14Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:14 crc kubenswrapper[4945]: E1206 06:54:14.225490 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.227785 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.227827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.227843 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.227864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.227877 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.331252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.331341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.331355 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.331375 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.331389 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.434837 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.435539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.435614 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.435687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.435769 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.539052 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.539114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.539133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.539155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.539172 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.621248 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:14 crc kubenswrapper[4945]: E1206 06:54:14.621510 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:54:14 crc kubenswrapper[4945]: E1206 06:54:14.621650 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs podName:d4326ced-ae97-4301-854b-222062d4764e nodeName:}" failed. No retries permitted until 2025-12-06 06:54:46.621620241 +0000 UTC m=+100.076481465 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs") pod "network-metrics-daemon-2xkp9" (UID: "d4326ced-ae97-4301-854b-222062d4764e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.642476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.642526 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.642538 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.642561 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.642581 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.745571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.745624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.745639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.745679 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.745694 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.849505 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.849952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.850090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.850329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.850519 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.953499 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.953930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.954059 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.954162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:14 crc kubenswrapper[4945]: I1206 06:54:14.954257 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:14Z","lastTransitionTime":"2025-12-06T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.057488 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.057530 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.057544 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.057562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.057574 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:15Z","lastTransitionTime":"2025-12-06T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.160780 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.161135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.161208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.161301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.161381 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:15Z","lastTransitionTime":"2025-12-06T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.264931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.264991 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.265002 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.265025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.265039 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:15Z","lastTransitionTime":"2025-12-06T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.368736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.368776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.368785 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.368801 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.368812 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:15Z","lastTransitionTime":"2025-12-06T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.472404 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.472505 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.472526 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.472556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.472580 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:15Z","lastTransitionTime":"2025-12-06T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.575835 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.575891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.575905 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.575928 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.575940 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:15Z","lastTransitionTime":"2025-12-06T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.679700 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.679766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.679791 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.679813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.679825 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:15Z","lastTransitionTime":"2025-12-06T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.783394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.783442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.783453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.783473 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.783485 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:15Z","lastTransitionTime":"2025-12-06T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.890066 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.890207 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.890234 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.890299 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.890326 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:15Z","lastTransitionTime":"2025-12-06T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.953135 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.953192 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:15 crc kubenswrapper[4945]: E1206 06:54:15.953372 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:15 crc kubenswrapper[4945]: E1206 06:54:15.953575 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.953601 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:15 crc kubenswrapper[4945]: E1206 06:54:15.953789 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.953925 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:15 crc kubenswrapper[4945]: E1206 06:54:15.954023 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.992881 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.992934 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.992947 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.992966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:15 crc kubenswrapper[4945]: I1206 06:54:15.992982 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:15Z","lastTransitionTime":"2025-12-06T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.097072 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.097134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.097146 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.097167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.097179 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:16Z","lastTransitionTime":"2025-12-06T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.199938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.199994 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.200011 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.200039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.200060 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:16Z","lastTransitionTime":"2025-12-06T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.302471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.302543 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.302562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.302586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.302605 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:16Z","lastTransitionTime":"2025-12-06T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.401273 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6v6gj_c9fc852c-2d5d-444c-b470-2728bb3dcdfd/kube-multus/0.log" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.401388 4945 generic.go:334] "Generic (PLEG): container finished" podID="c9fc852c-2d5d-444c-b470-2728bb3dcdfd" containerID="844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8" exitCode=1 Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.401441 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6v6gj" event={"ID":"c9fc852c-2d5d-444c-b470-2728bb3dcdfd","Type":"ContainerDied","Data":"844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.402111 4945 scope.go:117] "RemoveContainer" containerID="844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.404684 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.404742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.404753 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.404778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.404794 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:16Z","lastTransitionTime":"2025-12-06T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.421537 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.439224 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"2025-12-06T06:53:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d\\\\n2025-12-06T06:53:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d to /host/opt/cni/bin/\\\\n2025-12-06T06:53:30Z [verbose] multus-daemon started\\\\n2025-12-06T06:53:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T06:54:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.455298 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.471320 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.489024 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.505320 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.507123 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.507157 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.507169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.507190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.507204 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:16Z","lastTransitionTime":"2025-12-06T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.521957 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.535980 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.550995 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.573899 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.595186 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.609429 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.610422 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.610586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.610696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.610800 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.610913 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:16Z","lastTransitionTime":"2025-12-06T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.623316 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.646059 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"message\\\":\\\":57.859558 6576 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:57.859756 6576 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.859903 6576 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:57.859916 6576 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860016 6576 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.859968 6576 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 06:53:57.859992 6576 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.860260 6576 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860716 6576 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.659175 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.677013 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.691067 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4815516-b4c5-469b-8709-75608f08ed9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.703603 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d70cce2b-7da1-4c0d-be8d-cbe33328e93e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cc3d8f60c331dcab45117a8f5825236ba0b3b30939fd45727d2c2384d81964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.714354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.714777 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.714873 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.714967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.715056 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:16Z","lastTransitionTime":"2025-12-06T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.716170 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.817722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.817775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.817788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.817808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.817820 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:16Z","lastTransitionTime":"2025-12-06T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.920004 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.920048 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.920057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.920073 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.920092 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:16Z","lastTransitionTime":"2025-12-06T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.967530 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:16 crc kubenswrapper[4945]: I1206 06:54:16.990439 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4815516-b4c5-469b-8709-75608f08ed9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:16Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.006951 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d70cce2b-7da1-4c0d-be8d-cbe33328e93e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cc3d8f60c331dcab45117a8f5825236ba0b3b30939fd45727d2c2384d81964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.023092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.023158 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.023178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.023206 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.023226 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:17Z","lastTransitionTime":"2025-12-06T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.032039 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.059361 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"message\\\":\\\":57.859558 6576 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:57.859756 6576 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.859903 6576 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:57.859916 6576 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860016 6576 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.859968 6576 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 06:53:57.859992 6576 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.860260 6576 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860716 6576 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.080237 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.094659 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.107465 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"2025-12-06T06:53:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d\\\\n2025-12-06T06:53:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d to /host/opt/cni/bin/\\\\n2025-12-06T06:53:30Z [verbose] multus-daemon started\\\\n2025-12-06T06:53:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T06:54:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.119647 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.125527 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.125567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.125580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.125597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.125610 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:17Z","lastTransitionTime":"2025-12-06T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.131001 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.146495 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.159332 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.176868 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.199538 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.214211 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.228383 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.228437 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.228451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.228471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.228487 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:17Z","lastTransitionTime":"2025-12-06T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.229186 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.243708 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.256264 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.268413 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.331015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.331089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.331102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.331122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.331137 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:17Z","lastTransitionTime":"2025-12-06T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.408221 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6v6gj_c9fc852c-2d5d-444c-b470-2728bb3dcdfd/kube-multus/0.log" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.408336 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6v6gj" event={"ID":"c9fc852c-2d5d-444c-b470-2728bb3dcdfd","Type":"ContainerStarted","Data":"ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4"} Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.431610 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"message\\\":\\\":57.859558 6576 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:57.859756 6576 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.859903 6576 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:57.859916 6576 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860016 6576 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.859968 6576 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 06:53:57.859992 6576 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.860260 6576 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860716 6576 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.434408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.434452 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.434463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.434483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.434492 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:17Z","lastTransitionTime":"2025-12-06T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.449698 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.469085 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.485314 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4815516-b4c5-469b-8709-75608f08ed9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.499554 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d70cce2b-7da1-4c0d-be8d-cbe33328e93e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cc3d8f60c331dcab45117a8f5825236ba0b3b30939fd45727d2c2384d81964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.528401 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.538241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.538511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.538721 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.538886 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.539028 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:17Z","lastTransitionTime":"2025-12-06T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.544810 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.561995 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"2025-12-06T06:53:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d\\\\n2025-12-06T06:53:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d to /host/opt/cni/bin/\\\\n2025-12-06T06:53:30Z [verbose] multus-daemon started\\\\n2025-12-06T06:53:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T06:54:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:54:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.573838 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.587613 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.603590 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.618788 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.632331 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.642174 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.642226 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.642236 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.642254 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.642265 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:17Z","lastTransitionTime":"2025-12-06T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.643379 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.657255 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.689451 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.708523 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.723361 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.741363 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:17Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.746038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.746098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.746114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.746137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.746156 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:17Z","lastTransitionTime":"2025-12-06T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.848590 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.848640 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.848654 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.848674 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.848685 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:17Z","lastTransitionTime":"2025-12-06T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.952133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.952190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.952202 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.952225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.952225 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.952242 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.952317 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.952240 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:17Z","lastTransitionTime":"2025-12-06T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:17 crc kubenswrapper[4945]: E1206 06:54:17.952395 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:17 crc kubenswrapper[4945]: I1206 06:54:17.952513 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:17 crc kubenswrapper[4945]: E1206 06:54:17.952656 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:17 crc kubenswrapper[4945]: E1206 06:54:17.952738 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:17 crc kubenswrapper[4945]: E1206 06:54:17.952785 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.055596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.055652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.055664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.055682 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.055695 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:18Z","lastTransitionTime":"2025-12-06T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.159238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.159319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.159333 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.159350 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.159362 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:18Z","lastTransitionTime":"2025-12-06T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.262033 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.262144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.262154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.262176 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.262188 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:18Z","lastTransitionTime":"2025-12-06T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.364412 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.364471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.364498 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.364524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.364556 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:18Z","lastTransitionTime":"2025-12-06T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.467522 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.467565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.467573 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.467591 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.467600 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:18Z","lastTransitionTime":"2025-12-06T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.570335 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.570380 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.570395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.570413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.570424 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:18Z","lastTransitionTime":"2025-12-06T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.673136 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.673255 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.673353 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.673398 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.673425 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:18Z","lastTransitionTime":"2025-12-06T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.777511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.777588 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.777604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.777631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.777648 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:18Z","lastTransitionTime":"2025-12-06T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.881477 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.881525 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.881537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.881560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.881574 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:18Z","lastTransitionTime":"2025-12-06T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.984272 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.984402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.984418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.984440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:18 crc kubenswrapper[4945]: I1206 06:54:18.984462 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:18Z","lastTransitionTime":"2025-12-06T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.087571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.087615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.087626 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.087658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.087672 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:19Z","lastTransitionTime":"2025-12-06T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.190618 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.190687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.190705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.190731 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.190751 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:19Z","lastTransitionTime":"2025-12-06T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.293363 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.293421 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.293435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.293454 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.293468 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:19Z","lastTransitionTime":"2025-12-06T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.395853 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.395913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.395927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.395950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.395964 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:19Z","lastTransitionTime":"2025-12-06T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.499605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.499657 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.499668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.499687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.499698 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:19Z","lastTransitionTime":"2025-12-06T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.602085 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.602122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.602130 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.602146 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.602159 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:19Z","lastTransitionTime":"2025-12-06T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.704349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.704394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.704404 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.704421 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.704433 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:19Z","lastTransitionTime":"2025-12-06T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.807255 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.807328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.807352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.807373 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.807409 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:19Z","lastTransitionTime":"2025-12-06T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.910889 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.910938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.910949 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.910967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.910978 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:19Z","lastTransitionTime":"2025-12-06T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.952306 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.952360 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:19 crc kubenswrapper[4945]: E1206 06:54:19.952474 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.952514 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:19 crc kubenswrapper[4945]: I1206 06:54:19.952547 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:19 crc kubenswrapper[4945]: E1206 06:54:19.952648 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:19 crc kubenswrapper[4945]: E1206 06:54:19.952751 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:19 crc kubenswrapper[4945]: E1206 06:54:19.952816 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.013352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.013391 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.013402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.013419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.013432 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:20Z","lastTransitionTime":"2025-12-06T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.116320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.116361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.116372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.116388 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.116399 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:20Z","lastTransitionTime":"2025-12-06T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.218953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.218999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.219011 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.219030 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.219044 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:20Z","lastTransitionTime":"2025-12-06T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.322575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.322634 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.322653 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.322681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.322702 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:20Z","lastTransitionTime":"2025-12-06T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.424672 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.424704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.424713 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.424730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.424740 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:20Z","lastTransitionTime":"2025-12-06T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.527357 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.527408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.527420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.527441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.527454 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:20Z","lastTransitionTime":"2025-12-06T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.631455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.631518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.631531 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.631552 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.631581 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:20Z","lastTransitionTime":"2025-12-06T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.735015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.735074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.735084 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.735101 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.735115 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:20Z","lastTransitionTime":"2025-12-06T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.838884 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.838957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.838976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.839002 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.839032 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:20Z","lastTransitionTime":"2025-12-06T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.942317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.942358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.942368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.942384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:20 crc kubenswrapper[4945]: I1206 06:54:20.942395 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:20Z","lastTransitionTime":"2025-12-06T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.045581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.045677 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.045696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.045751 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.045770 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:21Z","lastTransitionTime":"2025-12-06T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.149047 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.149122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.149140 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.149166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.149184 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:21Z","lastTransitionTime":"2025-12-06T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.253671 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.253724 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.253764 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.253785 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.253803 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:21Z","lastTransitionTime":"2025-12-06T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.357071 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.357131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.357150 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.357177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.357195 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:21Z","lastTransitionTime":"2025-12-06T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.460006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.460053 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.460070 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.460088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.460103 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:21Z","lastTransitionTime":"2025-12-06T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.567485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.567717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.568335 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.568394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.568408 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:21Z","lastTransitionTime":"2025-12-06T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.672383 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.672480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.672495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.672516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.672532 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:21Z","lastTransitionTime":"2025-12-06T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.775253 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.775328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.775340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.775357 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.775370 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:21Z","lastTransitionTime":"2025-12-06T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.879007 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.879105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.879119 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.879137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.879148 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:21Z","lastTransitionTime":"2025-12-06T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.952707 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:21 crc kubenswrapper[4945]: E1206 06:54:21.952913 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.953095 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:21 crc kubenswrapper[4945]: E1206 06:54:21.953203 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.953338 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:21 crc kubenswrapper[4945]: E1206 06:54:21.953443 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.953600 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:21 crc kubenswrapper[4945]: E1206 06:54:21.953815 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.982009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.982091 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.982112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.982140 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:21 crc kubenswrapper[4945]: I1206 06:54:21.982160 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:21Z","lastTransitionTime":"2025-12-06T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.085349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.085396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.085408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.085428 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.085441 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:22Z","lastTransitionTime":"2025-12-06T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.188026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.188093 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.188103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.188120 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.188130 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:22Z","lastTransitionTime":"2025-12-06T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.292090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.292154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.292171 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.292194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.292212 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:22Z","lastTransitionTime":"2025-12-06T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.394983 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.395043 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.395054 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.395080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.395098 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:22Z","lastTransitionTime":"2025-12-06T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.497676 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.497723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.497738 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.497756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.497772 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:22Z","lastTransitionTime":"2025-12-06T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.600417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.600457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.600466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.600481 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.600493 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:22Z","lastTransitionTime":"2025-12-06T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.702982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.703041 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.703053 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.703071 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.703087 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:22Z","lastTransitionTime":"2025-12-06T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.806479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.806529 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.806538 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.806557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.806569 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:22Z","lastTransitionTime":"2025-12-06T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.910399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.910445 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.910457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.910472 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:22 crc kubenswrapper[4945]: I1206 06:54:22.910483 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:22Z","lastTransitionTime":"2025-12-06T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.013098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.013154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.013166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.013185 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.013200 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:23Z","lastTransitionTime":"2025-12-06T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.116317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.116720 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.116729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.116747 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.116758 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:23Z","lastTransitionTime":"2025-12-06T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.220533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.220608 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.220631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.220690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.220731 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:23Z","lastTransitionTime":"2025-12-06T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.323834 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.323909 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.323959 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.323991 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.324010 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:23Z","lastTransitionTime":"2025-12-06T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.427502 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.427552 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.427563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.427580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.427621 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:23Z","lastTransitionTime":"2025-12-06T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.531908 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.531995 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.532015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.532048 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.532073 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:23Z","lastTransitionTime":"2025-12-06T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.635805 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.635858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.635869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.635887 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.635899 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:23Z","lastTransitionTime":"2025-12-06T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.739832 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.739897 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.739914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.739940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.739960 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:23Z","lastTransitionTime":"2025-12-06T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.843932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.843995 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.844015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.844038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.844680 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:23Z","lastTransitionTime":"2025-12-06T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.948643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.948700 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.948710 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.948731 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.948747 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:23Z","lastTransitionTime":"2025-12-06T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.952765 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.952885 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:23 crc kubenswrapper[4945]: E1206 06:54:23.952930 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.952986 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:23 crc kubenswrapper[4945]: E1206 06:54:23.953064 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:23 crc kubenswrapper[4945]: E1206 06:54:23.953188 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:23 crc kubenswrapper[4945]: I1206 06:54:23.953450 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:23 crc kubenswrapper[4945]: E1206 06:54:23.953547 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.052532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.052613 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.052631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.052662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.052690 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.156485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.156567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.156585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.156612 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.156631 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.259673 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.259752 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.259767 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.259788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.259835 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.362719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.362757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.362767 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.362782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.362793 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.465532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.465577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.465594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.465616 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.465637 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.523580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.523640 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.523652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.523670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.523683 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: E1206 06:54:24.538088 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:24Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.542518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.542556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.542570 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.542589 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.542602 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: E1206 06:54:24.560020 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:24Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.564501 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.564657 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.564760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.564863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.564961 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: E1206 06:54:24.580069 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:24Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.585502 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.585575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.585595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.585622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.585643 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: E1206 06:54:24.600681 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:24Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.607414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.607473 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.607488 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.607510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.607526 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: E1206 06:54:24.628603 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:24Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:24 crc kubenswrapper[4945]: E1206 06:54:24.628862 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.631698 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.631760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.631774 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.631801 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.631816 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.735545 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.735658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.735723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.735760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.735828 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.839745 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.839823 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.839847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.839878 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.839905 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.943783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.943888 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.943913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.943945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:24 crc kubenswrapper[4945]: I1206 06:54:24.943969 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:24Z","lastTransitionTime":"2025-12-06T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.047850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.047910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.047925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.047951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.047969 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:25Z","lastTransitionTime":"2025-12-06T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.151084 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.151166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.151185 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.151212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.151235 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:25Z","lastTransitionTime":"2025-12-06T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.254631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.254737 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.254758 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.254783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.254804 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:25Z","lastTransitionTime":"2025-12-06T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.357931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.357983 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.358000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.358023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.358038 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:25Z","lastTransitionTime":"2025-12-06T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.467026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.467678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.467901 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.468094 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.468222 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:25Z","lastTransitionTime":"2025-12-06T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.571999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.572080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.572104 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.572139 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.572163 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:25Z","lastTransitionTime":"2025-12-06T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.675680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.675746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.675767 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.675790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.675804 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:25Z","lastTransitionTime":"2025-12-06T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.778345 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.778387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.778397 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.778413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.778426 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:25Z","lastTransitionTime":"2025-12-06T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.881700 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.881751 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.881764 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.881783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.881793 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:25Z","lastTransitionTime":"2025-12-06T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.952436 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.952506 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.952442 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.952571 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:25 crc kubenswrapper[4945]: E1206 06:54:25.952679 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:25 crc kubenswrapper[4945]: E1206 06:54:25.952875 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:25 crc kubenswrapper[4945]: E1206 06:54:25.953004 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:25 crc kubenswrapper[4945]: E1206 06:54:25.953124 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.985183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.985231 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.985244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.985260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:25 crc kubenswrapper[4945]: I1206 06:54:25.985295 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:25Z","lastTransitionTime":"2025-12-06T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.088689 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.088758 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.088779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.088808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.088828 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:26Z","lastTransitionTime":"2025-12-06T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.192433 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.192511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.192533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.192562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.192589 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:26Z","lastTransitionTime":"2025-12-06T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.295881 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.295969 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.295986 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.296006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.296046 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:26Z","lastTransitionTime":"2025-12-06T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.399567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.399630 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.399643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.399664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.399678 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:26Z","lastTransitionTime":"2025-12-06T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.502690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.502770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.502784 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.502803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.502819 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:26Z","lastTransitionTime":"2025-12-06T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.606139 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.606202 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.606220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.606243 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.606257 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:26Z","lastTransitionTime":"2025-12-06T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.709452 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.709504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.709514 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.709534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.709549 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:26Z","lastTransitionTime":"2025-12-06T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.812563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.812615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.812625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.812644 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.812655 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:26Z","lastTransitionTime":"2025-12-06T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.916909 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.916965 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.916974 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.916994 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.917006 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:26Z","lastTransitionTime":"2025-12-06T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.953847 4945 scope.go:117] "RemoveContainer" containerID="7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.971640 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:26Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.987765 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4815516-b4c5-469b-8709-75608f08ed9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:26Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:26 crc kubenswrapper[4945]: I1206 06:54:26.999532 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d70cce2b-7da1-4c0d-be8d-cbe33328e93e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cc3d8f60c331dcab45117a8f5825236ba0b3b30939fd45727d2c2384d81964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:26Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.015957 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.020424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.020573 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.020602 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.020626 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.020658 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:27Z","lastTransitionTime":"2025-12-06T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.048657 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"message\\\":\\\":57.859558 6576 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:57.859756 6576 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.859903 6576 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:57.859916 6576 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860016 6576 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.859968 6576 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 06:53:57.859992 6576 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.860260 6576 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860716 6576 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.065958 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.081916 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.097956 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"2025-12-06T06:53:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d\\\\n2025-12-06T06:53:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d to /host/opt/cni/bin/\\\\n2025-12-06T06:53:30Z [verbose] multus-daemon started\\\\n2025-12-06T06:53:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T06:54:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:54:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.113231 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.123890 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.123957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.123968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.123988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.124000 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:27Z","lastTransitionTime":"2025-12-06T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.128091 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.145818 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.163424 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.183063 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.206814 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.227552 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.227602 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.227618 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.227639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.227653 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:27Z","lastTransitionTime":"2025-12-06T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.228572 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.244647 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.260491 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.270486 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.281800 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:27Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.330462 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.330516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.330528 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.330550 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.330568 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:27Z","lastTransitionTime":"2025-12-06T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.433548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.433621 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.433642 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.433676 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.433695 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:27Z","lastTransitionTime":"2025-12-06T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.537238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.537319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.537330 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.537351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.537363 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:27Z","lastTransitionTime":"2025-12-06T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.640772 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.640835 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.640845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.640865 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.640879 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:27Z","lastTransitionTime":"2025-12-06T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.744531 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.744629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.744651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.744686 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.744729 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:27Z","lastTransitionTime":"2025-12-06T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.848535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.848584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.848595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.848610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.848619 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:27Z","lastTransitionTime":"2025-12-06T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.951688 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.951732 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.951743 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.951761 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.951774 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:27Z","lastTransitionTime":"2025-12-06T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.952482 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.952503 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.952524 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:27 crc kubenswrapper[4945]: I1206 06:54:27.952552 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:27 crc kubenswrapper[4945]: E1206 06:54:27.952725 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:27 crc kubenswrapper[4945]: E1206 06:54:27.952864 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:27 crc kubenswrapper[4945]: E1206 06:54:27.952898 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:27 crc kubenswrapper[4945]: E1206 06:54:27.952958 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.055130 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.055194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.055208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.055234 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.055250 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:28Z","lastTransitionTime":"2025-12-06T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.159045 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.159104 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.159123 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.159153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.159173 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:28Z","lastTransitionTime":"2025-12-06T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.261937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.261996 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.262008 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.262029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.262041 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:28Z","lastTransitionTime":"2025-12-06T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.364928 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.364982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.364996 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.365015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.365032 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:28Z","lastTransitionTime":"2025-12-06T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.452539 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/2.log" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.456034 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47"} Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.467240 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.467318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.467333 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.467358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.467372 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:28Z","lastTransitionTime":"2025-12-06T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.571176 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.571233 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.571243 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.571262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.571288 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:28Z","lastTransitionTime":"2025-12-06T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.674560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.674619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.674634 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.674652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.674662 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:28Z","lastTransitionTime":"2025-12-06T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.777775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.777837 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.777853 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.777877 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.777890 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:28Z","lastTransitionTime":"2025-12-06T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.879898 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.879944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.879953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.879970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.879992 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:28Z","lastTransitionTime":"2025-12-06T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.982247 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.982315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.982327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.982347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:28 crc kubenswrapper[4945]: I1206 06:54:28.982371 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:28Z","lastTransitionTime":"2025-12-06T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.085717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.085759 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.085771 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.085788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.085801 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:29Z","lastTransitionTime":"2025-12-06T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.188367 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.188449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.188475 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.188508 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.188529 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:29Z","lastTransitionTime":"2025-12-06T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.292087 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.292160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.292175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.292197 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.292215 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:29Z","lastTransitionTime":"2025-12-06T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.394942 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.395000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.395015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.395040 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.395050 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:29Z","lastTransitionTime":"2025-12-06T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.461561 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/3.log" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.462267 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/2.log" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.465504 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47" exitCode=1 Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.465575 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.465634 4945 scope.go:117] "RemoveContainer" containerID="7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.466752 4945 scope.go:117] "RemoveContainer" containerID="31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47" Dec 06 06:54:29 crc kubenswrapper[4945]: E1206 06:54:29.467184 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.484584 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.498821 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.498887 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.498900 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.498921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.498963 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:29Z","lastTransitionTime":"2025-12-06T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.500013 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.518306 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.533169 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.546300 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.565937 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.581717 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4815516-b4c5-469b-8709-75608f08ed9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.595928 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d70cce2b-7da1-4c0d-be8d-cbe33328e93e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cc3d8f60c331dcab45117a8f5825236ba0b3b30939fd45727d2c2384d81964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.601142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.601196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.601216 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.601241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.601258 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:29Z","lastTransitionTime":"2025-12-06T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.612379 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.637397 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d9d19df7047e0a4e526e3b38198ff40dca86f824ec8e2ef8d91932881f5ab97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"message\\\":\\\":57.859558 6576 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 06:53:57.859756 6576 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.859903 6576 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 06:53:57.859916 6576 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860016 6576 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.859968 6576 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 06:53:57.859992 6576 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 06:53:57.860260 6576 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 06:53:57.860716 6576 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:29Z\\\",\\\"message\\\":\\\"e{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1206 06:54:29.331780 6958 services_controller.go:360] Finished syncing service cluster-version-operator on namespace openshift-cluster-version for network=default : 2.935124ms\\\\nI1206 06:54:29.331807 6958 services_controller.go:356] Processing sync for service openshift-machine-api/cluster-autoscaler-operator for network=default\\\\nF1206 06:54:29.331882 6958 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not add\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:54:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.656044 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.676189 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.692592 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"2025-12-06T06:53:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d\\\\n2025-12-06T06:53:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d to /host/opt/cni/bin/\\\\n2025-12-06T06:53:30Z [verbose] multus-daemon started\\\\n2025-12-06T06:53:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T06:54:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:54:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.704012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.704065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.704079 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.704096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.704111 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:29Z","lastTransitionTime":"2025-12-06T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.710153 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.724351 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.738927 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.757584 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.776600 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.794033 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:29Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.806956 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.807028 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.807038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.807057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.807068 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:29Z","lastTransitionTime":"2025-12-06T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.910275 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.910394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.910412 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.910446 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.910509 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:29Z","lastTransitionTime":"2025-12-06T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.952854 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.952854 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.952878 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:29 crc kubenswrapper[4945]: I1206 06:54:29.953041 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:29 crc kubenswrapper[4945]: E1206 06:54:29.953124 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:29 crc kubenswrapper[4945]: E1206 06:54:29.953322 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:29 crc kubenswrapper[4945]: E1206 06:54:29.953441 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:29 crc kubenswrapper[4945]: E1206 06:54:29.953532 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.016513 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.016571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.016584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.016607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.017934 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:30Z","lastTransitionTime":"2025-12-06T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.120858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.120914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.120925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.120943 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.120955 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:30Z","lastTransitionTime":"2025-12-06T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.224186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.224249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.224262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.224305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.224318 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:30Z","lastTransitionTime":"2025-12-06T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.327861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.327911 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.327927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.327951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.327967 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:30Z","lastTransitionTime":"2025-12-06T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.430845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.430908 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.430927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.430945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.430960 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:30Z","lastTransitionTime":"2025-12-06T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.471978 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/3.log" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.476990 4945 scope.go:117] "RemoveContainer" containerID="31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47" Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.477306 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.496239 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.512240 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4815516-b4c5-469b-8709-75608f08ed9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.525489 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d70cce2b-7da1-4c0d-be8d-cbe33328e93e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cc3d8f60c331dcab45117a8f5825236ba0b3b30939fd45727d2c2384d81964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.533523 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.533571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.533581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.533598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.533610 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:30Z","lastTransitionTime":"2025-12-06T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.542438 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.563838 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:29Z\\\",\\\"message\\\":\\\"e{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1206 06:54:29.331780 6958 services_controller.go:360] Finished syncing service cluster-version-operator on namespace openshift-cluster-version for network=default : 2.935124ms\\\\nI1206 06:54:29.331807 6958 services_controller.go:356] Processing sync for service openshift-machine-api/cluster-autoscaler-operator for network=default\\\\nF1206 06:54:29.331882 6958 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not add\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:54:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.577936 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.590567 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.603074 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"2025-12-06T06:53:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d\\\\n2025-12-06T06:53:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d to /host/opt/cni/bin/\\\\n2025-12-06T06:53:30Z [verbose] multus-daemon started\\\\n2025-12-06T06:53:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T06:54:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:54:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.611985 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.612141 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.612263 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.612235217 +0000 UTC m=+148.067096261 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.615560 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.629143 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.635827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.635884 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.635899 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.635921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.635933 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:30Z","lastTransitionTime":"2025-12-06T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.643458 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.657251 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.672382 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.694782 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.710513 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.712687 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.712898 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.712957 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.712912206 +0000 UTC m=+148.167773250 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.713041 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.713110 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.71309451 +0000 UTC m=+148.167955554 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.713039 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.713215 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.713248 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.713302 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.713354 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.713343917 +0000 UTC m=+148.168205221 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.713353 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.713468 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.713488 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.713502 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:54:30 crc kubenswrapper[4945]: E1206 06:54:30.713546 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.713537672 +0000 UTC m=+148.168398716 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.722733 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.735921 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.738261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.738319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.738331 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.738354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.738369 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:30Z","lastTransitionTime":"2025-12-06T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.746516 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.759897 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:30Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.840893 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.840938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.840948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.840968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.840979 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:30Z","lastTransitionTime":"2025-12-06T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.944041 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.944081 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.944090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.944106 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:30 crc kubenswrapper[4945]: I1206 06:54:30.944117 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:30Z","lastTransitionTime":"2025-12-06T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.047485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.047543 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.047555 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.047576 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.047589 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:31Z","lastTransitionTime":"2025-12-06T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.150942 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.151047 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.151074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.151112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.151139 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:31Z","lastTransitionTime":"2025-12-06T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.254609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.254656 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.254667 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.254686 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.254702 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:31Z","lastTransitionTime":"2025-12-06T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.363474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.363523 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.363536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.363556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.363568 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:31Z","lastTransitionTime":"2025-12-06T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.466848 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.466918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.466932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.466951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.466968 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:31Z","lastTransitionTime":"2025-12-06T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.570317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.570395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.570414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.570440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.570459 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:31Z","lastTransitionTime":"2025-12-06T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.675214 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.675257 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.675267 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.675297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.675309 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:31Z","lastTransitionTime":"2025-12-06T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.778565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.778622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.778632 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.778651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.778661 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:31Z","lastTransitionTime":"2025-12-06T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.882757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.882818 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.882832 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.882854 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.882874 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:31Z","lastTransitionTime":"2025-12-06T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.952878 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.952948 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.952949 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:31 crc kubenswrapper[4945]: E1206 06:54:31.953046 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.953114 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:31 crc kubenswrapper[4945]: E1206 06:54:31.953195 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:31 crc kubenswrapper[4945]: E1206 06:54:31.953254 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:31 crc kubenswrapper[4945]: E1206 06:54:31.953341 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.986174 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.986235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.986251 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.986272 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:31 crc kubenswrapper[4945]: I1206 06:54:31.986303 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:31Z","lastTransitionTime":"2025-12-06T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.089055 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.089116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.089129 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.089151 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.089166 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:32Z","lastTransitionTime":"2025-12-06T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.192193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.192236 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.192244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.192262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.192287 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:32Z","lastTransitionTime":"2025-12-06T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.294917 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.294957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.294967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.294984 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.294995 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:32Z","lastTransitionTime":"2025-12-06T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.398127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.398213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.398235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.398264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.398317 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:32Z","lastTransitionTime":"2025-12-06T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.500880 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.500921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.500930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.500946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.500958 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:32Z","lastTransitionTime":"2025-12-06T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.603678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.603738 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.603752 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.603775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.603787 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:32Z","lastTransitionTime":"2025-12-06T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.706740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.706791 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.706802 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.706822 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.706835 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:32Z","lastTransitionTime":"2025-12-06T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.810063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.810125 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.810135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.810152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.810163 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:32Z","lastTransitionTime":"2025-12-06T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.913327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.913376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.913387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.913406 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:32 crc kubenswrapper[4945]: I1206 06:54:32.913423 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:32Z","lastTransitionTime":"2025-12-06T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.016674 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.016754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.016790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.016813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.016826 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:33Z","lastTransitionTime":"2025-12-06T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.120579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.120652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.120673 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.120698 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.120713 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:33Z","lastTransitionTime":"2025-12-06T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.222808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.222853 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.222863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.222879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.222893 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:33Z","lastTransitionTime":"2025-12-06T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.325974 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.326026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.326040 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.326060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.326073 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:33Z","lastTransitionTime":"2025-12-06T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.427808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.427879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.427888 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.427904 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.427915 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:33Z","lastTransitionTime":"2025-12-06T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.530300 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.530343 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.530352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.530368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.530379 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:33Z","lastTransitionTime":"2025-12-06T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.633322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.633371 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.633383 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.633401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.633413 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:33Z","lastTransitionTime":"2025-12-06T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.737486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.737549 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.737563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.737584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.737599 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:33Z","lastTransitionTime":"2025-12-06T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.842957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.843029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.843050 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.843080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.843106 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:33Z","lastTransitionTime":"2025-12-06T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.946248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.946340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.946356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.946374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.946412 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:33Z","lastTransitionTime":"2025-12-06T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.952346 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.952420 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.952426 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:33 crc kubenswrapper[4945]: I1206 06:54:33.952486 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:33 crc kubenswrapper[4945]: E1206 06:54:33.952616 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:33 crc kubenswrapper[4945]: E1206 06:54:33.952746 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:33 crc kubenswrapper[4945]: E1206 06:54:33.952860 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:33 crc kubenswrapper[4945]: E1206 06:54:33.953011 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.050399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.050473 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.050492 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.050524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.050547 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.153922 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.154001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.154027 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.154062 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.154086 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.257091 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.257173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.257193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.257222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.257243 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.359958 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.360007 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.360018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.360042 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.360055 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.462800 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.462893 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.462916 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.462950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.462973 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.566467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.566556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.566582 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.566619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.566644 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.670119 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.670193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.670235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.670312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.670342 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.773721 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.773782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.773802 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.773831 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.773849 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.877579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.877660 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.877684 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.877721 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.877746 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.927170 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.927324 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.927351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.927381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.927400 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: E1206 06:54:34.952704 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.958881 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.958938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.958950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.958968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.958984 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:34 crc kubenswrapper[4945]: E1206 06:54:34.979433 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.985360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.985463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.985482 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.985508 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:34 crc kubenswrapper[4945]: I1206 06:54:34.985524 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:34Z","lastTransitionTime":"2025-12-06T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: E1206 06:54:35.001129 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:34Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.006075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.006122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.006136 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.006153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.006166 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: E1206 06:54:35.020164 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.024670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.024715 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.024729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.024751 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.024763 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: E1206 06:54:35.037146 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:35Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:35 crc kubenswrapper[4945]: E1206 06:54:35.037272 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.039035 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.039078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.039094 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.039113 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.039125 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.141589 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.141648 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.141659 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.141681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.141693 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.244875 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.244940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.244953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.245027 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.245045 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.348431 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.348536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.348553 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.348579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.348595 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.451221 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.451264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.451286 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.451307 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.451328 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.561206 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.561295 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.561308 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.561329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.561341 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.664268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.664398 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.664420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.664445 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.664463 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.767722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.767795 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.767819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.767847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.767871 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.870029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.870110 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.870135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.870164 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.870190 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.952342 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.952397 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.952349 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.952366 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:35 crc kubenswrapper[4945]: E1206 06:54:35.952521 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:35 crc kubenswrapper[4945]: E1206 06:54:35.952696 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:35 crc kubenswrapper[4945]: E1206 06:54:35.952856 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:35 crc kubenswrapper[4945]: E1206 06:54:35.952912 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.972902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.972954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.972967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.972988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:35 crc kubenswrapper[4945]: I1206 06:54:35.973004 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:35Z","lastTransitionTime":"2025-12-06T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.076015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.076070 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.076088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.076109 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.076125 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:36Z","lastTransitionTime":"2025-12-06T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.179646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.179701 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.179713 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.179732 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.179744 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:36Z","lastTransitionTime":"2025-12-06T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.282415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.282456 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.282467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.282485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.282499 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:36Z","lastTransitionTime":"2025-12-06T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.385116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.385188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.385206 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.385229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.385246 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:36Z","lastTransitionTime":"2025-12-06T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.488237 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.488354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.488378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.488411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.488436 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:36Z","lastTransitionTime":"2025-12-06T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.592708 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.592763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.592780 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.592804 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.592817 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:36Z","lastTransitionTime":"2025-12-06T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.695600 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.695655 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.695670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.695691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.695708 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:36Z","lastTransitionTime":"2025-12-06T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.798976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.799023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.799034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.799055 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.799069 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:36Z","lastTransitionTime":"2025-12-06T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.902178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.902235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.902248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.902272 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.902306 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:36Z","lastTransitionTime":"2025-12-06T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.967308 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.981331 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:36 crc kubenswrapper[4945]: I1206 06:54:36.995704 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:36Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.005207 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.005311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.005325 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.005346 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.005359 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:37Z","lastTransitionTime":"2025-12-06T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.024891 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.043222 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.061834 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.076133 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.096340 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:29Z\\\",\\\"message\\\":\\\"e{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1206 06:54:29.331780 6958 services_controller.go:360] Finished syncing service cluster-version-operator on namespace openshift-cluster-version for network=default : 2.935124ms\\\\nI1206 06:54:29.331807 6958 services_controller.go:356] Processing sync for service openshift-machine-api/cluster-autoscaler-operator for network=default\\\\nF1206 06:54:29.331882 6958 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not add\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:54:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.108228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.108355 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.108382 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.108421 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.108453 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:37Z","lastTransitionTime":"2025-12-06T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.111398 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.128019 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.144522 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4815516-b4c5-469b-8709-75608f08ed9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.155856 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d70cce2b-7da1-4c0d-be8d-cbe33328e93e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cc3d8f60c331dcab45117a8f5825236ba0b3b30939fd45727d2c2384d81964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.169889 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.185848 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.200542 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"2025-12-06T06:53:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d\\\\n2025-12-06T06:53:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d to /host/opt/cni/bin/\\\\n2025-12-06T06:53:30Z [verbose] multus-daemon started\\\\n2025-12-06T06:53:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T06:54:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:54:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.216177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.216238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.216250 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.216321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.216353 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:37Z","lastTransitionTime":"2025-12-06T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.217296 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.232196 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.244761 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.257551 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:37Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.320203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.320310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.320329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.320358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.320377 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:37Z","lastTransitionTime":"2025-12-06T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.422697 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.422755 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.422771 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.422794 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.422808 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:37Z","lastTransitionTime":"2025-12-06T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.526769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.526823 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.526859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.526876 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.526887 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:37Z","lastTransitionTime":"2025-12-06T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.630713 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.630794 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.630815 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.630872 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.630893 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:37Z","lastTransitionTime":"2025-12-06T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.735077 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.735143 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.735162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.735187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.735206 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:37Z","lastTransitionTime":"2025-12-06T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.838696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.838753 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.838769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.838790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.838835 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:37Z","lastTransitionTime":"2025-12-06T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.942084 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.942128 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.942138 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.942155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.942169 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:37Z","lastTransitionTime":"2025-12-06T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.952868 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:37 crc kubenswrapper[4945]: E1206 06:54:37.953021 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.953690 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.953789 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:37 crc kubenswrapper[4945]: E1206 06:54:37.953881 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:37 crc kubenswrapper[4945]: I1206 06:54:37.954031 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:37 crc kubenswrapper[4945]: E1206 06:54:37.954096 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:37 crc kubenswrapper[4945]: E1206 06:54:37.954262 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.045081 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.045124 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.045143 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.045164 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.045178 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:38Z","lastTransitionTime":"2025-12-06T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.148964 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.149025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.149039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.149061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.149076 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:38Z","lastTransitionTime":"2025-12-06T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.251337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.251386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.251398 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.251414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.251427 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:38Z","lastTransitionTime":"2025-12-06T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.354965 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.355039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.355059 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.355090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.355113 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:38Z","lastTransitionTime":"2025-12-06T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.459060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.459135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.459149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.459171 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.459186 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:38Z","lastTransitionTime":"2025-12-06T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.562403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.562461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.562474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.562496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.562510 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:38Z","lastTransitionTime":"2025-12-06T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.664961 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.665003 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.665014 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.665029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.665039 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:38Z","lastTransitionTime":"2025-12-06T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.768731 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.768789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.768800 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.768817 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.768830 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:38Z","lastTransitionTime":"2025-12-06T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.871246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.871310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.871323 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.871340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.871352 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:38Z","lastTransitionTime":"2025-12-06T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.974221 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.974758 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.974774 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.974790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:38 crc kubenswrapper[4945]: I1206 06:54:38.974805 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:38Z","lastTransitionTime":"2025-12-06T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.079395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.079491 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.079517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.079553 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.079577 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:39Z","lastTransitionTime":"2025-12-06T06:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.182316 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.182376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.182391 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.182415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.182432 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:39Z","lastTransitionTime":"2025-12-06T06:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.285250 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.285327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.285339 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.285358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.285370 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:39Z","lastTransitionTime":"2025-12-06T06:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.388196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.388325 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.388357 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.388390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.388423 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:39Z","lastTransitionTime":"2025-12-06T06:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.492030 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.492077 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.492089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.492108 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.492121 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:39Z","lastTransitionTime":"2025-12-06T06:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.595483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.595546 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.595564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.595589 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.595608 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:39Z","lastTransitionTime":"2025-12-06T06:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.698118 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.698185 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.698204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.698233 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.698253 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:39Z","lastTransitionTime":"2025-12-06T06:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.801508 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.801558 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.801645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.801666 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.801680 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:39Z","lastTransitionTime":"2025-12-06T06:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.904681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.904752 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.904766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.904787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.904801 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:39Z","lastTransitionTime":"2025-12-06T06:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.952701 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.952732 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.952897 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:39 crc kubenswrapper[4945]: E1206 06:54:39.952884 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:39 crc kubenswrapper[4945]: I1206 06:54:39.952980 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:39 crc kubenswrapper[4945]: E1206 06:54:39.953069 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:39 crc kubenswrapper[4945]: E1206 06:54:39.953180 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:39 crc kubenswrapper[4945]: E1206 06:54:39.953228 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.007270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.007339 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.007358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.007378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.007390 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:40Z","lastTransitionTime":"2025-12-06T06:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.110818 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.110881 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.110898 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.110972 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.110991 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:40Z","lastTransitionTime":"2025-12-06T06:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.213820 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.213870 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.213883 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.213902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.213916 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:40Z","lastTransitionTime":"2025-12-06T06:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.317008 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.317071 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.317090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.317107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.317119 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:40Z","lastTransitionTime":"2025-12-06T06:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.420356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.420404 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.420418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.420438 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.420454 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:40Z","lastTransitionTime":"2025-12-06T06:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.522506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.522551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.522565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.522583 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.522594 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:40Z","lastTransitionTime":"2025-12-06T06:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.625588 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.625652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.625664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.625682 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.625694 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:40Z","lastTransitionTime":"2025-12-06T06:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.728586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.728633 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.728645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.728665 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.728678 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:40Z","lastTransitionTime":"2025-12-06T06:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.831800 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.831870 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.831885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.831923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.831937 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:40Z","lastTransitionTime":"2025-12-06T06:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.934763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.934830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.934847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.934870 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:40 crc kubenswrapper[4945]: I1206 06:54:40.934885 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:40Z","lastTransitionTime":"2025-12-06T06:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.037111 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.037151 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.037160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.037175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.037186 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:41Z","lastTransitionTime":"2025-12-06T06:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.140360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.140422 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.140432 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.140448 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.140459 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:41Z","lastTransitionTime":"2025-12-06T06:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.243386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.243432 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.243443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.243461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.243474 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:41Z","lastTransitionTime":"2025-12-06T06:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.347181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.347316 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.347345 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.347378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.347400 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:41Z","lastTransitionTime":"2025-12-06T06:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.450455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.450502 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.450516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.450532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.450545 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:41Z","lastTransitionTime":"2025-12-06T06:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.553718 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.553758 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.553768 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.553785 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.553797 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:41Z","lastTransitionTime":"2025-12-06T06:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.655977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.656023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.656035 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.656052 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.656062 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:41Z","lastTransitionTime":"2025-12-06T06:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.760067 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.760135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.760150 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.760175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.760195 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:41Z","lastTransitionTime":"2025-12-06T06:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.863998 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.864058 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.864070 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.864089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.864101 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:41Z","lastTransitionTime":"2025-12-06T06:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.952517 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.952646 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.952693 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:41 crc kubenswrapper[4945]: E1206 06:54:41.952941 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.953018 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:41 crc kubenswrapper[4945]: E1206 06:54:41.953176 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:41 crc kubenswrapper[4945]: E1206 06:54:41.953233 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:41 crc kubenswrapper[4945]: E1206 06:54:41.954197 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.955034 4945 scope.go:117] "RemoveContainer" containerID="31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47" Dec 06 06:54:41 crc kubenswrapper[4945]: E1206 06:54:41.955406 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.967134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.967190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.967207 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.967273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:41 crc kubenswrapper[4945]: I1206 06:54:41.967332 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:41Z","lastTransitionTime":"2025-12-06T06:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.070319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.070364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.070374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.070392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.070404 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:42Z","lastTransitionTime":"2025-12-06T06:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.172665 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.172709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.172718 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.172735 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.172748 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:42Z","lastTransitionTime":"2025-12-06T06:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.275425 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.275846 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.275859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.275880 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.275898 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:42Z","lastTransitionTime":"2025-12-06T06:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.378788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.378852 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.378872 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.378896 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.378913 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:42Z","lastTransitionTime":"2025-12-06T06:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.482534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.482581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.482592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.482611 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.482623 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:42Z","lastTransitionTime":"2025-12-06T06:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.585212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.585308 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.585324 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.585345 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.585361 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:42Z","lastTransitionTime":"2025-12-06T06:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.687949 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.688005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.688014 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.688032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.688042 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:42Z","lastTransitionTime":"2025-12-06T06:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.791073 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.791133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.791144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.791162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.791173 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:42Z","lastTransitionTime":"2025-12-06T06:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.893780 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.893856 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.893878 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.893910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.893932 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:42Z","lastTransitionTime":"2025-12-06T06:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.996554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.996617 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.996627 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.996643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:42 crc kubenswrapper[4945]: I1206 06:54:42.996683 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:42Z","lastTransitionTime":"2025-12-06T06:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.100114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.100201 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.100222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.100252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.100271 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:43Z","lastTransitionTime":"2025-12-06T06:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.204202 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.204262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.204306 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.204347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.204363 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:43Z","lastTransitionTime":"2025-12-06T06:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.254627 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.255505 4945 scope.go:117] "RemoveContainer" containerID="31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47" Dec 06 06:54:43 crc kubenswrapper[4945]: E1206 06:54:43.255695 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.306794 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.306865 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.306880 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.306904 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.306918 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:43Z","lastTransitionTime":"2025-12-06T06:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.410875 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.410927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.410938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.410961 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.410973 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:43Z","lastTransitionTime":"2025-12-06T06:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.514333 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.514386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.514418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.514459 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.514481 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:43Z","lastTransitionTime":"2025-12-06T06:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.618139 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.618183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.618193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.618213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.618224 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:43Z","lastTransitionTime":"2025-12-06T06:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.721554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.721678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.721696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.721720 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.721736 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:43Z","lastTransitionTime":"2025-12-06T06:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.825008 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.825051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.825070 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.825089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.825099 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:43Z","lastTransitionTime":"2025-12-06T06:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.928410 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.928460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.928471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.928490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.928502 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:43Z","lastTransitionTime":"2025-12-06T06:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.952825 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.952910 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.952987 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:43 crc kubenswrapper[4945]: E1206 06:54:43.952995 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:43 crc kubenswrapper[4945]: I1206 06:54:43.953061 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:43 crc kubenswrapper[4945]: E1206 06:54:43.953147 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:43 crc kubenswrapper[4945]: E1206 06:54:43.953230 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:43 crc kubenswrapper[4945]: E1206 06:54:43.953330 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.031165 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.031223 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.031237 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.031261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.031279 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:44Z","lastTransitionTime":"2025-12-06T06:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.134018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.134076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.134086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.134102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.134113 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:44Z","lastTransitionTime":"2025-12-06T06:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.237680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.237730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.237740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.237756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.237767 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:44Z","lastTransitionTime":"2025-12-06T06:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.341123 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.341170 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.341179 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.341197 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.341209 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:44Z","lastTransitionTime":"2025-12-06T06:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.444637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.444693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.444704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.444722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.444735 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:44Z","lastTransitionTime":"2025-12-06T06:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.548179 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.548229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.548242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.548266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.548314 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:44Z","lastTransitionTime":"2025-12-06T06:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.651961 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.652020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.652034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.652059 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.652077 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:44Z","lastTransitionTime":"2025-12-06T06:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.755889 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.755982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.756006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.756037 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.756060 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:44Z","lastTransitionTime":"2025-12-06T06:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.859328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.859381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.859394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.859415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.859427 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:44Z","lastTransitionTime":"2025-12-06T06:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.962260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.962319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.962332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.962350 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:44 crc kubenswrapper[4945]: I1206 06:54:44.962378 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:44Z","lastTransitionTime":"2025-12-06T06:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.065205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.065337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.065363 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.065397 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.065418 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.169263 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.169392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.169431 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.169468 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.169492 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.253418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.253474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.253487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.253504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.253516 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: E1206 06:54:45.269614 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:45Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.275810 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.275885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.275907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.275930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.275948 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: E1206 06:54:45.291008 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:45Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.295061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.295116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.295125 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.295141 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.295173 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: E1206 06:54:45.311527 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:45Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.316107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.316189 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.316203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.316232 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.316248 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: E1206 06:54:45.330601 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:45Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.335155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.335216 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.335228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.335251 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.335345 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: E1206 06:54:45.349273 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"15bb9f62-3356-4dc4-bade-82cf13a5e0d3\\\",\\\"systemUUID\\\":\\\"4af87963-9c01-463b-9f2d-189a21242b1b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:45Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:45 crc kubenswrapper[4945]: E1206 06:54:45.349433 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.351799 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.351846 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.351857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.351876 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.351889 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.454955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.455015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.455029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.455050 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.455066 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.558008 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.558067 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.558080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.558102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.558118 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.660746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.660785 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.660796 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.660811 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.660821 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.764575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.764632 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.764650 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.764676 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.764696 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.876436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.876491 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.876504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.876526 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.876542 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.952337 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.952381 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.952404 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.952444 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:45 crc kubenswrapper[4945]: E1206 06:54:45.952529 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:45 crc kubenswrapper[4945]: E1206 06:54:45.952667 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:45 crc kubenswrapper[4945]: E1206 06:54:45.952844 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:45 crc kubenswrapper[4945]: E1206 06:54:45.952912 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.979635 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.979713 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.979727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.980113 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:45 crc kubenswrapper[4945]: I1206 06:54:45.980151 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:45Z","lastTransitionTime":"2025-12-06T06:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.084106 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.084166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.084177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.084195 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.084208 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:46Z","lastTransitionTime":"2025-12-06T06:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.187696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.187770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.187789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.187818 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.187840 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:46Z","lastTransitionTime":"2025-12-06T06:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.290862 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.290938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.290957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.290984 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.290998 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:46Z","lastTransitionTime":"2025-12-06T06:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.393997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.394047 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.394059 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.394077 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.394091 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:46Z","lastTransitionTime":"2025-12-06T06:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.496740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.496795 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.496871 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.496898 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.496916 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:46Z","lastTransitionTime":"2025-12-06T06:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.600368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.600441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.600452 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.600469 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.600480 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:46Z","lastTransitionTime":"2025-12-06T06:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.697236 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:46 crc kubenswrapper[4945]: E1206 06:54:46.697452 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:54:46 crc kubenswrapper[4945]: E1206 06:54:46.697544 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs podName:d4326ced-ae97-4301-854b-222062d4764e nodeName:}" failed. No retries permitted until 2025-12-06 06:55:50.697522228 +0000 UTC m=+164.152383282 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs") pod "network-metrics-daemon-2xkp9" (UID: "d4326ced-ae97-4301-854b-222062d4764e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.703092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.703144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.703156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.703172 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.703183 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:46Z","lastTransitionTime":"2025-12-06T06:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.805540 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.805590 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.805601 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.805619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.805885 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:46Z","lastTransitionTime":"2025-12-06T06:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.910012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.910076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.910089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.910112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.910154 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:46Z","lastTransitionTime":"2025-12-06T06:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.971719 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b03f9e01e29288e48923a83666ada0f2eac8e5255a17573cca0efbacf1d0b230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:46Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:46 crc kubenswrapper[4945]: I1206 06:54:46.988866 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6v6gj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9fc852c-2d5d-444c-b470-2728bb3dcdfd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:54:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:16Z\\\",\\\"message\\\":\\\"2025-12-06T06:53:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d\\\\n2025-12-06T06:53:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4d4c34b7-6195-4ddb-8329-3e0a0766105d to /host/opt/cni/bin/\\\\n2025-12-06T06:53:30Z [verbose] multus-daemon started\\\\n2025-12-06T06:53:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T06:54:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:54:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-grvtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6v6gj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:46Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.003314 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-np9zn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d67cbd5-c8f0-499f-9255-01ad8dc4071f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a60a11e605a3715662963548106cd07a69d949d05547105303b95cee408dc303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kkxsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-np9zn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:46Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.012408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.012478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.012492 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.012510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.012522 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:47Z","lastTransitionTime":"2025-12-06T06:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.017256 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4326ced-ae97-4301-854b-222062d4764e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b58mz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2xkp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.033356 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b924ff1-dd26-4e47-b90c-144eae88b737\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T06:53:20Z\\\",\\\"message\\\":\\\"W1206 06:53:10.146256 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 06:53:10.146792 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765003990 cert, and key in /tmp/serving-cert-1721764549/serving-signer.crt, /tmp/serving-cert-1721764549/serving-signer.key\\\\nI1206 06:53:10.354533 1 observer_polling.go:159] Starting file observer\\\\nW1206 06:53:10.356601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 06:53:10.356941 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 06:53:10.358305 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1721764549/tls.crt::/tmp/serving-cert-1721764549/tls.key\\\\\\\"\\\\nF1206 06:53:20.721717 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.050599 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db19c229e861dcd93b0e8c85c33cfdaaf6dc74d68d7da2fb79294bdf0fbd993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f1921890a2d3c9d794aa09556da1b2ffd4a422c2c43b29dacff471dffd9839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.069979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.091170 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44908297-a79a-48d0-b1a8-4bbe953f8904\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a96690102d5d906870a2c4cec21fc80668fe440e4ea0faf665b932ba95b07cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7a6cec9fb8bf99bbb104263d76e9c519ef3dd34718eeae134b75ec9164e8a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://318470041b6ace121f2d3e9eeb623526c9749ed1a39662c5b927ad50f79c7ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d733c27172c49b8a242ce8f6a85100db35acaa71e4ad8a6f737ab7263f233676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a60eba3ea202eb655186e9d67e93a99281aec0bec0373464e1720da33de4833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a808b1901fd376d65bdbdcc62147c3f887a3ba659f1d0c60e9442cda3848b333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd919aa7354e1478dc295bf980f976fc68fec8970afdd81ff7aa9b553bd97a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc618dbbefc3d070334de64d245ef2a47d879d4855985e3620cb51bf6b6c436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.104454 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831e8b7633a554715e185efcfb25157d7c3d6fd973a5f65cc286bcdfd9268260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.115417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.115456 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.115468 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.115487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.115500 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:47Z","lastTransitionTime":"2025-12-06T06:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.119195 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.133001 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.144654 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hw42h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60d17d4-5e9c-4e43-b584-54e7187712b7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://336a5d4ffe98a0ca33c9e85db07d7c7266ae3baa8e0c6e2435f2256baaf11088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vlvjz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hw42h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.161436 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f56279eb-fe4e-4bd2-bccd-d39958e923ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfbb2eb7530781ad84a9287e8607027bf97a6f22d7582bdab1248b29ac4e41e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-57tvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.179212 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c3531e-88e8-4af7-9870-733be5c04f07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3d348a7db4e1e98869ed150eb00f8437ebcb48b8c27846979fc9f0ba867da2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a256ffc04261d3f70aaf8b94b91c2add296f8cb1263761802b1ab741dbbe903\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9da95d1a03bdb57f9ffcf445a4e10e99c38c13f07329dd8598d658564330337\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.197555 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4815516-b4c5-469b-8709-75608f08ed9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b20635e5d1e0ce49dd9caa63fe7aa7d7af6884531655bcb22854a1553f0ee22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c25de8a31f76742357311042399dc2644caf30e83fa636535b7d28c95e96ad3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f8eb9cf427f504186ebcad531a895f5ab3b75eaa80ebbfe85bd1660386cad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b47afdb3729cdde8f0fa7aa6ba4624715ee452fb648863cf88efdbce57eb1cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.212210 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d70cce2b-7da1-4c0d-be8d-cbe33328e93e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cc3d8f60c331dcab45117a8f5825236ba0b3b30939fd45727d2c2384d81964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b90f10d572579a4197b39dfdc0282b34b556d08449c48137434432d6ab44a6ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.218663 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.218706 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.218760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.218786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.218802 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:47Z","lastTransitionTime":"2025-12-06T06:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.229195 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jrv66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f379f13c-33b1-4e56-81d5-9244ceb6cabd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21507de3b7a0ca7afe04ebfcab8929606546643b9df959a74b2ea84ab58a4996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1570f14a049c54752f8dd1427d9143b12e1bb3d17c4a6c121527ee0b73dadb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c2bcb46c22e485024c00a903051a5372caf6bd0f9678a21307aa7beea417120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5534d0da54f02bac8fd9482186306adf41f617ad4697c0f793f4c3a169299e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd83f564c373cac228b0be9b55b23510463e3a438412ef5633134cf7c4fc6e75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d059eda39a351edc9397ffaae5f0c1bc4aa10611508c09e0e3e4be813aa64e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7888e8bc33b6e65a21173300c1b11e015b3cad9b7593f454dc95389f58e13459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g659f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jrv66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.253463 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T06:54:29Z\\\",\\\"message\\\":\\\"e{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1206 06:54:29.331780 6958 services_controller.go:360] Finished syncing service cluster-version-operator on namespace openshift-cluster-version for network=default : 2.935124ms\\\\nI1206 06:54:29.331807 6958 services_controller.go:356] Processing sync for service openshift-machine-api/cluster-autoscaler-operator for network=default\\\\nF1206 06:54:29.331882 6958 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not add\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T06:54:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T06:53:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T06:53:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qh8t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mlqp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.268829 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"680cc0bb-d74c-42de-8dcc-0690426fdfa5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T06:53:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f4e86bf32ec86c50a354b10c2026574af00cb80ccbd7fa751bcdd519f0a0b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff04109371550a424725cc82a3b1764b98ebb7e3a3d6c6a41049a0c624b1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T06:53:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T06:53:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tprpz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T06:54:47Z is after 2025-08-24T17:21:41Z" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.323415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.323466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.323483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.323512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.323533 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:47Z","lastTransitionTime":"2025-12-06T06:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.426474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.426554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.426580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.426612 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.426633 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:47Z","lastTransitionTime":"2025-12-06T06:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.530117 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.530187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.530203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.530227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.530245 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:47Z","lastTransitionTime":"2025-12-06T06:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.632971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.633035 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.633056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.633080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.633097 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:47Z","lastTransitionTime":"2025-12-06T06:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.736687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.736772 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.736795 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.736824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.736845 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:47Z","lastTransitionTime":"2025-12-06T06:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.840495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.840577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.840589 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.840611 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.840625 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:47Z","lastTransitionTime":"2025-12-06T06:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.943399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.943447 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.943457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.943475 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.943489 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:47Z","lastTransitionTime":"2025-12-06T06:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.952992 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.953140 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.952992 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:47 crc kubenswrapper[4945]: E1206 06:54:47.953189 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:47 crc kubenswrapper[4945]: I1206 06:54:47.953000 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:47 crc kubenswrapper[4945]: E1206 06:54:47.953357 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:47 crc kubenswrapper[4945]: E1206 06:54:47.953479 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:47 crc kubenswrapper[4945]: E1206 06:54:47.953583 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.045851 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.045902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.045911 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.045927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.045939 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:48Z","lastTransitionTime":"2025-12-06T06:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.149410 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.149457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.149465 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.149493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.149507 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:48Z","lastTransitionTime":"2025-12-06T06:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.251974 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.252040 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.252060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.252082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.252094 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:48Z","lastTransitionTime":"2025-12-06T06:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.354780 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.355134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.355247 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.355407 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.355528 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:48Z","lastTransitionTime":"2025-12-06T06:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.458615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.458654 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.458664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.458682 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.458693 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:48Z","lastTransitionTime":"2025-12-06T06:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.561714 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.561779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.561794 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.561811 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.561823 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:48Z","lastTransitionTime":"2025-12-06T06:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.664425 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.664466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.664474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.664487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.664497 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:48Z","lastTransitionTime":"2025-12-06T06:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.767849 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.767910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.767929 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.767953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.767972 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:48Z","lastTransitionTime":"2025-12-06T06:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.871046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.871096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.871109 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.871130 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.871141 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:48Z","lastTransitionTime":"2025-12-06T06:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.974038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.974080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.974088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.974102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:48 crc kubenswrapper[4945]: I1206 06:54:48.974113 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:48Z","lastTransitionTime":"2025-12-06T06:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.077094 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.077157 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.077183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.077204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.077219 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:49Z","lastTransitionTime":"2025-12-06T06:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.180555 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.180957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.180977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.181010 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.181027 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:49Z","lastTransitionTime":"2025-12-06T06:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.286727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.286775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.286788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.286807 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.286818 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:49Z","lastTransitionTime":"2025-12-06T06:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.389765 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.389814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.389831 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.389861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.389875 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:49Z","lastTransitionTime":"2025-12-06T06:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.492393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.492436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.492448 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.492466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.492478 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:49Z","lastTransitionTime":"2025-12-06T06:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.595976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.596025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.596035 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.596054 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.596066 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:49Z","lastTransitionTime":"2025-12-06T06:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.699470 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.699549 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.699566 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.699590 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.699609 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:49Z","lastTransitionTime":"2025-12-06T06:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.802225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.802264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.802273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.802303 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.802315 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:49Z","lastTransitionTime":"2025-12-06T06:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.906086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.906124 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.906133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.906148 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.906157 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:49Z","lastTransitionTime":"2025-12-06T06:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.952206 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:49 crc kubenswrapper[4945]: E1206 06:54:49.952372 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.952589 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.952637 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:49 crc kubenswrapper[4945]: I1206 06:54:49.952637 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:49 crc kubenswrapper[4945]: E1206 06:54:49.952821 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:49 crc kubenswrapper[4945]: E1206 06:54:49.952655 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:49 crc kubenswrapper[4945]: E1206 06:54:49.952884 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.009771 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.009830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.009853 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.009878 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.009898 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:50Z","lastTransitionTime":"2025-12-06T06:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.113325 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.113381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.113403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.113429 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.113450 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:50Z","lastTransitionTime":"2025-12-06T06:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.216387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.216458 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.216480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.216505 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.216527 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:50Z","lastTransitionTime":"2025-12-06T06:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.320368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.320438 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.320451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.320471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.320485 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:50Z","lastTransitionTime":"2025-12-06T06:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.423321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.423367 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.423377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.423394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.423404 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:50Z","lastTransitionTime":"2025-12-06T06:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.525772 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.525835 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.525846 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.525863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.525877 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:50Z","lastTransitionTime":"2025-12-06T06:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.628347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.628401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.628411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.628427 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.628439 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:50Z","lastTransitionTime":"2025-12-06T06:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.730914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.730972 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.730984 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.731006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.731021 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:50Z","lastTransitionTime":"2025-12-06T06:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.834235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.834372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.834392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.834440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.834456 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:50Z","lastTransitionTime":"2025-12-06T06:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.936907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.937003 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.937020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.937044 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:50 crc kubenswrapper[4945]: I1206 06:54:50.937056 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:50Z","lastTransitionTime":"2025-12-06T06:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.040435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.040486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.040495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.040516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.040527 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:51Z","lastTransitionTime":"2025-12-06T06:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.144574 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.144637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.144653 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.144684 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.144705 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:51Z","lastTransitionTime":"2025-12-06T06:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.248242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.248338 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.248358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.248384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.248401 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:51Z","lastTransitionTime":"2025-12-06T06:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.350625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.350697 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.350720 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.350749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.350770 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:51Z","lastTransitionTime":"2025-12-06T06:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.453331 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.453416 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.453443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.453481 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.453506 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:51Z","lastTransitionTime":"2025-12-06T06:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.555390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.555491 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.555507 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.555529 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.555544 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:51Z","lastTransitionTime":"2025-12-06T06:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.658065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.658112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.658128 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.658147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.658160 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:51Z","lastTransitionTime":"2025-12-06T06:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.761073 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.761122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.761135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.761154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.761166 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:51Z","lastTransitionTime":"2025-12-06T06:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.863719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.863820 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.863841 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.863866 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.863889 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:51Z","lastTransitionTime":"2025-12-06T06:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.953359 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.953440 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:51 crc kubenswrapper[4945]: E1206 06:54:51.953538 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.953634 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:51 crc kubenswrapper[4945]: E1206 06:54:51.954649 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:51 crc kubenswrapper[4945]: E1206 06:54:51.953644 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.955611 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:51 crc kubenswrapper[4945]: E1206 06:54:51.962008 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.966092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.966133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.966142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.966158 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:51 crc kubenswrapper[4945]: I1206 06:54:51.966169 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:51Z","lastTransitionTime":"2025-12-06T06:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.069474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.069550 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.069575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.069608 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.069632 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:52Z","lastTransitionTime":"2025-12-06T06:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.172135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.172183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.172195 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.172220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.172233 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:52Z","lastTransitionTime":"2025-12-06T06:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.275753 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.275807 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.275816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.275835 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.275846 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:52Z","lastTransitionTime":"2025-12-06T06:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.378200 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.378304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.378320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.378343 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.378358 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:52Z","lastTransitionTime":"2025-12-06T06:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.481067 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.481127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.481139 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.481160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.481178 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:52Z","lastTransitionTime":"2025-12-06T06:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.584395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.584454 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.584465 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.584483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.584497 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:52Z","lastTransitionTime":"2025-12-06T06:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.687556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.687658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.687674 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.687704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.687726 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:52Z","lastTransitionTime":"2025-12-06T06:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.791499 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.791546 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.791563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.791583 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.791597 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:52Z","lastTransitionTime":"2025-12-06T06:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.894758 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.894802 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.894812 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.894828 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.894840 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:52Z","lastTransitionTime":"2025-12-06T06:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.997348 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.997410 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.997420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.997437 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:52 crc kubenswrapper[4945]: I1206 06:54:52.997468 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:52Z","lastTransitionTime":"2025-12-06T06:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.100806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.100872 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.100885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.100902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.100913 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:53Z","lastTransitionTime":"2025-12-06T06:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.203587 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.203639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.203649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.203664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.203674 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:53Z","lastTransitionTime":"2025-12-06T06:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.307101 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.307154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.307168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.307187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.307201 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:53Z","lastTransitionTime":"2025-12-06T06:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.410462 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.410532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.410547 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.410571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.410586 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:53Z","lastTransitionTime":"2025-12-06T06:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.513479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.513538 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.513548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.513563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.513572 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:53Z","lastTransitionTime":"2025-12-06T06:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.616645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.616750 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.616769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.616798 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.616815 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:53Z","lastTransitionTime":"2025-12-06T06:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.719806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.719885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.719914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.719947 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.719974 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:53Z","lastTransitionTime":"2025-12-06T06:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.822755 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.822855 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.822885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.822917 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.822939 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:53Z","lastTransitionTime":"2025-12-06T06:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.925698 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.925739 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.925748 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.925763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.925773 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:53Z","lastTransitionTime":"2025-12-06T06:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.952326 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:53 crc kubenswrapper[4945]: E1206 06:54:53.952495 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.952339 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.952359 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:53 crc kubenswrapper[4945]: I1206 06:54:53.952339 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:53 crc kubenswrapper[4945]: E1206 06:54:53.952645 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:53 crc kubenswrapper[4945]: E1206 06:54:53.952816 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:53 crc kubenswrapper[4945]: E1206 06:54:53.952917 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.027662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.027717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.027727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.027746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.027756 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:54Z","lastTransitionTime":"2025-12-06T06:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.130542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.130589 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.130598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.130615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.130627 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:54Z","lastTransitionTime":"2025-12-06T06:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.233219 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.233294 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.233312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.233340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.233356 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:54Z","lastTransitionTime":"2025-12-06T06:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.336231 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.336293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.336306 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.336329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.336342 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:54Z","lastTransitionTime":"2025-12-06T06:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.439185 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.439260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.439316 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.439342 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.439355 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:54Z","lastTransitionTime":"2025-12-06T06:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.543747 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.543817 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.543840 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.543865 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.543880 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:54Z","lastTransitionTime":"2025-12-06T06:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.654194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.654669 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.654824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.654940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.655260 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:54Z","lastTransitionTime":"2025-12-06T06:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.759455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.759524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.759536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.759559 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.759574 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:54Z","lastTransitionTime":"2025-12-06T06:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.862640 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.862710 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.862724 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.862749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.862762 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:54Z","lastTransitionTime":"2025-12-06T06:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.966337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.966387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.966396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.966417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:54 crc kubenswrapper[4945]: I1206 06:54:54.966439 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:54Z","lastTransitionTime":"2025-12-06T06:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.069369 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.069414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.069425 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.069439 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.069449 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:55Z","lastTransitionTime":"2025-12-06T06:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.171949 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.171998 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.172009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.172029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.172043 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:55Z","lastTransitionTime":"2025-12-06T06:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.274467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.274518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.274535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.274565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.274578 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:55Z","lastTransitionTime":"2025-12-06T06:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.377364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.377410 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.377420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.377436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.377448 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:55Z","lastTransitionTime":"2025-12-06T06:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.415510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.415564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.415575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.415597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.415609 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T06:54:55Z","lastTransitionTime":"2025-12-06T06:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.479503 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h"] Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.480056 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.482529 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.482860 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.482903 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.483455 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.565051 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=89.565024835 podStartE2EDuration="1m29.565024835s" podCreationTimestamp="2025-12-06 06:53:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.549531823 +0000 UTC m=+109.004392867" watchObservedRunningTime="2025-12-06 06:54:55.565024835 +0000 UTC m=+109.019885879" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.601450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.601532 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.601561 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.601587 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.601645 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.609415 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-hw42h" podStartSLOduration=87.609391008 podStartE2EDuration="1m27.609391008s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.609271345 +0000 UTC m=+109.064132409" watchObservedRunningTime="2025-12-06 06:54:55.609391008 +0000 UTC m=+109.064252092" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.623365 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podStartSLOduration=87.623335921 podStartE2EDuration="1m27.623335921s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.62330041 +0000 UTC m=+109.078161464" watchObservedRunningTime="2025-12-06 06:54:55.623335921 +0000 UTC m=+109.078196975" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.650902 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=89.650879108 podStartE2EDuration="1m29.650879108s" podCreationTimestamp="2025-12-06 06:53:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.650001476 +0000 UTC m=+109.104862540" watchObservedRunningTime="2025-12-06 06:54:55.650879108 +0000 UTC m=+109.105740162" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.676892 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=57.676864576 podStartE2EDuration="57.676864576s" podCreationTimestamp="2025-12-06 06:53:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.663945139 +0000 UTC m=+109.118806213" watchObservedRunningTime="2025-12-06 06:54:55.676864576 +0000 UTC m=+109.131725620" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.695491 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=42.695465057 podStartE2EDuration="42.695465057s" podCreationTimestamp="2025-12-06 06:54:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.676791894 +0000 UTC m=+109.131652948" watchObservedRunningTime="2025-12-06 06:54:55.695465057 +0000 UTC m=+109.150326111" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.696370 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-jrv66" podStartSLOduration=87.69636434 podStartE2EDuration="1m27.69636434s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.696068792 +0000 UTC m=+109.150929846" watchObservedRunningTime="2025-12-06 06:54:55.69636434 +0000 UTC m=+109.151225394" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.702562 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.702622 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.702642 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.702667 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.702705 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.702775 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.702842 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.703824 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.713712 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.723673 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82897cfc-0e52-43f9-9ce7-bd4eba92fa28-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wlm4h\" (UID: \"82897cfc-0e52-43f9-9ce7-bd4eba92fa28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.741540 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tprpz" podStartSLOduration=86.741510612 podStartE2EDuration="1m26.741510612s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.740103427 +0000 UTC m=+109.194964511" watchObservedRunningTime="2025-12-06 06:54:55.741510612 +0000 UTC m=+109.196371696" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.753757 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=85.753718731 podStartE2EDuration="1m25.753718731s" podCreationTimestamp="2025-12-06 06:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.753114626 +0000 UTC m=+109.207975720" watchObservedRunningTime="2025-12-06 06:54:55.753718731 +0000 UTC m=+109.208579775" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.772020 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-6v6gj" podStartSLOduration=87.771986064 podStartE2EDuration="1m27.771986064s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.770923677 +0000 UTC m=+109.225784731" watchObservedRunningTime="2025-12-06 06:54:55.771986064 +0000 UTC m=+109.226847148" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.783989 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-np9zn" podStartSLOduration=87.783970347 podStartE2EDuration="1m27.783970347s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:55.783459284 +0000 UTC m=+109.238320328" watchObservedRunningTime="2025-12-06 06:54:55.783970347 +0000 UTC m=+109.238831431" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.797119 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.953353 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.953474 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.953378 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:55 crc kubenswrapper[4945]: E1206 06:54:55.953535 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:55 crc kubenswrapper[4945]: E1206 06:54:55.953625 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:55 crc kubenswrapper[4945]: I1206 06:54:55.953652 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:55 crc kubenswrapper[4945]: E1206 06:54:55.953729 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:55 crc kubenswrapper[4945]: E1206 06:54:55.953802 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:56 crc kubenswrapper[4945]: I1206 06:54:56.574975 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" event={"ID":"82897cfc-0e52-43f9-9ce7-bd4eba92fa28","Type":"ContainerStarted","Data":"7eb1feac57187661a301fd63b8faca6477f0d4c4626b5cc268035a8826c45951"} Dec 06 06:54:56 crc kubenswrapper[4945]: I1206 06:54:56.575036 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" event={"ID":"82897cfc-0e52-43f9-9ce7-bd4eba92fa28","Type":"ContainerStarted","Data":"ad7ce12b91bc47167463fc64b2bab2a864bff60839bc9c0845bfdc8e198e348a"} Dec 06 06:54:56 crc kubenswrapper[4945]: I1206 06:54:56.592195 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wlm4h" podStartSLOduration=88.592172506 podStartE2EDuration="1m28.592172506s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:54:56.591075058 +0000 UTC m=+110.045936102" watchObservedRunningTime="2025-12-06 06:54:56.592172506 +0000 UTC m=+110.047033550" Dec 06 06:54:57 crc kubenswrapper[4945]: I1206 06:54:57.952452 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:57 crc kubenswrapper[4945]: I1206 06:54:57.952470 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:57 crc kubenswrapper[4945]: I1206 06:54:57.952497 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:57 crc kubenswrapper[4945]: I1206 06:54:57.952627 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:57 crc kubenswrapper[4945]: E1206 06:54:57.953224 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:54:57 crc kubenswrapper[4945]: E1206 06:54:57.953387 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:57 crc kubenswrapper[4945]: E1206 06:54:57.953476 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:57 crc kubenswrapper[4945]: E1206 06:54:57.953696 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:57 crc kubenswrapper[4945]: I1206 06:54:57.953869 4945 scope.go:117] "RemoveContainer" containerID="31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47" Dec 06 06:54:57 crc kubenswrapper[4945]: E1206 06:54:57.954226 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mlqp5_openshift-ovn-kubernetes(4d69e33c-ba3b-4fca-a77a-a098db9d24c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" Dec 06 06:54:59 crc kubenswrapper[4945]: I1206 06:54:59.952496 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:54:59 crc kubenswrapper[4945]: I1206 06:54:59.952564 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:54:59 crc kubenswrapper[4945]: I1206 06:54:59.952579 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:54:59 crc kubenswrapper[4945]: I1206 06:54:59.952626 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:54:59 crc kubenswrapper[4945]: E1206 06:54:59.952648 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:54:59 crc kubenswrapper[4945]: E1206 06:54:59.952710 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:54:59 crc kubenswrapper[4945]: E1206 06:54:59.952802 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:54:59 crc kubenswrapper[4945]: E1206 06:54:59.953076 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:01 crc kubenswrapper[4945]: I1206 06:55:01.953090 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:01 crc kubenswrapper[4945]: I1206 06:55:01.953216 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:01 crc kubenswrapper[4945]: E1206 06:55:01.953334 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:01 crc kubenswrapper[4945]: E1206 06:55:01.953417 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:01 crc kubenswrapper[4945]: I1206 06:55:01.953432 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:01 crc kubenswrapper[4945]: E1206 06:55:01.953545 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:01 crc kubenswrapper[4945]: I1206 06:55:01.953623 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:01 crc kubenswrapper[4945]: E1206 06:55:01.953695 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:02 crc kubenswrapper[4945]: I1206 06:55:02.596441 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6v6gj_c9fc852c-2d5d-444c-b470-2728bb3dcdfd/kube-multus/1.log" Dec 06 06:55:02 crc kubenswrapper[4945]: I1206 06:55:02.597343 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6v6gj_c9fc852c-2d5d-444c-b470-2728bb3dcdfd/kube-multus/0.log" Dec 06 06:55:02 crc kubenswrapper[4945]: I1206 06:55:02.597391 4945 generic.go:334] "Generic (PLEG): container finished" podID="c9fc852c-2d5d-444c-b470-2728bb3dcdfd" containerID="ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4" exitCode=1 Dec 06 06:55:02 crc kubenswrapper[4945]: I1206 06:55:02.597430 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6v6gj" event={"ID":"c9fc852c-2d5d-444c-b470-2728bb3dcdfd","Type":"ContainerDied","Data":"ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4"} Dec 06 06:55:02 crc kubenswrapper[4945]: I1206 06:55:02.597476 4945 scope.go:117] "RemoveContainer" containerID="844f2b4af3f1b2c71b2cfd611d91d5aeaba57f12daf88e40eea85b89424ee0d8" Dec 06 06:55:02 crc kubenswrapper[4945]: I1206 06:55:02.598064 4945 scope.go:117] "RemoveContainer" containerID="ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4" Dec 06 06:55:02 crc kubenswrapper[4945]: E1206 06:55:02.598312 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-6v6gj_openshift-multus(c9fc852c-2d5d-444c-b470-2728bb3dcdfd)\"" pod="openshift-multus/multus-6v6gj" podUID="c9fc852c-2d5d-444c-b470-2728bb3dcdfd" Dec 06 06:55:03 crc kubenswrapper[4945]: I1206 06:55:03.602939 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6v6gj_c9fc852c-2d5d-444c-b470-2728bb3dcdfd/kube-multus/1.log" Dec 06 06:55:03 crc kubenswrapper[4945]: I1206 06:55:03.953121 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:03 crc kubenswrapper[4945]: I1206 06:55:03.953147 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:03 crc kubenswrapper[4945]: E1206 06:55:03.953247 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:03 crc kubenswrapper[4945]: E1206 06:55:03.953407 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:03 crc kubenswrapper[4945]: I1206 06:55:03.953403 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:03 crc kubenswrapper[4945]: E1206 06:55:03.953521 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:03 crc kubenswrapper[4945]: I1206 06:55:03.953955 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:03 crc kubenswrapper[4945]: E1206 06:55:03.954174 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:05 crc kubenswrapper[4945]: I1206 06:55:05.952644 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:05 crc kubenswrapper[4945]: I1206 06:55:05.952661 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:05 crc kubenswrapper[4945]: E1206 06:55:05.952807 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:05 crc kubenswrapper[4945]: I1206 06:55:05.952775 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:05 crc kubenswrapper[4945]: I1206 06:55:05.952644 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:05 crc kubenswrapper[4945]: E1206 06:55:05.952899 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:05 crc kubenswrapper[4945]: E1206 06:55:05.953125 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:05 crc kubenswrapper[4945]: E1206 06:55:05.953163 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:06 crc kubenswrapper[4945]: E1206 06:55:06.948954 4945 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 06 06:55:07 crc kubenswrapper[4945]: E1206 06:55:07.055580 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:55:07 crc kubenswrapper[4945]: I1206 06:55:07.953372 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:07 crc kubenswrapper[4945]: I1206 06:55:07.953463 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:07 crc kubenswrapper[4945]: I1206 06:55:07.953506 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:07 crc kubenswrapper[4945]: I1206 06:55:07.953408 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:07 crc kubenswrapper[4945]: E1206 06:55:07.953667 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:07 crc kubenswrapper[4945]: E1206 06:55:07.953746 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:07 crc kubenswrapper[4945]: E1206 06:55:07.953859 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:07 crc kubenswrapper[4945]: E1206 06:55:07.954047 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:09 crc kubenswrapper[4945]: I1206 06:55:09.952869 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:09 crc kubenswrapper[4945]: I1206 06:55:09.952997 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:09 crc kubenswrapper[4945]: E1206 06:55:09.953065 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:09 crc kubenswrapper[4945]: I1206 06:55:09.953089 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:09 crc kubenswrapper[4945]: I1206 06:55:09.953503 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:09 crc kubenswrapper[4945]: E1206 06:55:09.953659 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:09 crc kubenswrapper[4945]: E1206 06:55:09.953861 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:09 crc kubenswrapper[4945]: E1206 06:55:09.953897 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:09 crc kubenswrapper[4945]: I1206 06:55:09.953963 4945 scope.go:117] "RemoveContainer" containerID="31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47" Dec 06 06:55:10 crc kubenswrapper[4945]: I1206 06:55:10.630569 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/3.log" Dec 06 06:55:10 crc kubenswrapper[4945]: I1206 06:55:10.634356 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerStarted","Data":"d7d68c21448dc3cdea139233e91a91ca289f3040b698a3d09456305979ab0585"} Dec 06 06:55:10 crc kubenswrapper[4945]: I1206 06:55:10.635531 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:55:10 crc kubenswrapper[4945]: I1206 06:55:10.680097 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podStartSLOduration=102.680077719 podStartE2EDuration="1m42.680077719s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:10.67774468 +0000 UTC m=+124.132605734" watchObservedRunningTime="2025-12-06 06:55:10.680077719 +0000 UTC m=+124.134938763" Dec 06 06:55:10 crc kubenswrapper[4945]: I1206 06:55:10.902863 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2xkp9"] Dec 06 06:55:10 crc kubenswrapper[4945]: I1206 06:55:10.903066 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:10 crc kubenswrapper[4945]: E1206 06:55:10.903209 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:11 crc kubenswrapper[4945]: I1206 06:55:11.952597 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:11 crc kubenswrapper[4945]: I1206 06:55:11.952619 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:11 crc kubenswrapper[4945]: E1206 06:55:11.952879 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:11 crc kubenswrapper[4945]: I1206 06:55:11.953052 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:11 crc kubenswrapper[4945]: E1206 06:55:11.953230 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:11 crc kubenswrapper[4945]: E1206 06:55:11.953555 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:12 crc kubenswrapper[4945]: E1206 06:55:12.058335 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:55:12 crc kubenswrapper[4945]: I1206 06:55:12.952912 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:12 crc kubenswrapper[4945]: E1206 06:55:12.953100 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:13 crc kubenswrapper[4945]: I1206 06:55:13.952488 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:13 crc kubenswrapper[4945]: I1206 06:55:13.952573 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:13 crc kubenswrapper[4945]: I1206 06:55:13.952452 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:13 crc kubenswrapper[4945]: E1206 06:55:13.952763 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:13 crc kubenswrapper[4945]: E1206 06:55:13.952879 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:13 crc kubenswrapper[4945]: E1206 06:55:13.953016 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:14 crc kubenswrapper[4945]: I1206 06:55:14.952509 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:14 crc kubenswrapper[4945]: E1206 06:55:14.952672 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:15 crc kubenswrapper[4945]: I1206 06:55:15.952988 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:15 crc kubenswrapper[4945]: E1206 06:55:15.953499 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:15 crc kubenswrapper[4945]: I1206 06:55:15.953043 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:15 crc kubenswrapper[4945]: I1206 06:55:15.953043 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:15 crc kubenswrapper[4945]: E1206 06:55:15.953687 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:15 crc kubenswrapper[4945]: I1206 06:55:15.953731 4945 scope.go:117] "RemoveContainer" containerID="ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4" Dec 06 06:55:15 crc kubenswrapper[4945]: E1206 06:55:15.953746 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:16 crc kubenswrapper[4945]: I1206 06:55:16.952938 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:16 crc kubenswrapper[4945]: E1206 06:55:16.954233 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:17 crc kubenswrapper[4945]: E1206 06:55:17.059159 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:55:17 crc kubenswrapper[4945]: I1206 06:55:17.952387 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:17 crc kubenswrapper[4945]: I1206 06:55:17.952426 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:17 crc kubenswrapper[4945]: I1206 06:55:17.952388 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:17 crc kubenswrapper[4945]: E1206 06:55:17.952606 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:17 crc kubenswrapper[4945]: E1206 06:55:17.952845 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:17 crc kubenswrapper[4945]: E1206 06:55:17.952997 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:18 crc kubenswrapper[4945]: I1206 06:55:18.668231 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6v6gj_c9fc852c-2d5d-444c-b470-2728bb3dcdfd/kube-multus/1.log" Dec 06 06:55:18 crc kubenswrapper[4945]: I1206 06:55:18.668371 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6v6gj" event={"ID":"c9fc852c-2d5d-444c-b470-2728bb3dcdfd","Type":"ContainerStarted","Data":"f45f16e4b54f593b1b291ff367b569587c385448071b36e1380fe53e9c3f6411"} Dec 06 06:55:18 crc kubenswrapper[4945]: I1206 06:55:18.952630 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:18 crc kubenswrapper[4945]: E1206 06:55:18.952876 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:19 crc kubenswrapper[4945]: I1206 06:55:19.953145 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:19 crc kubenswrapper[4945]: I1206 06:55:19.953146 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:19 crc kubenswrapper[4945]: E1206 06:55:19.953372 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:19 crc kubenswrapper[4945]: I1206 06:55:19.953176 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:19 crc kubenswrapper[4945]: E1206 06:55:19.953479 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:19 crc kubenswrapper[4945]: E1206 06:55:19.953607 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:20 crc kubenswrapper[4945]: I1206 06:55:20.952457 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:20 crc kubenswrapper[4945]: E1206 06:55:20.952695 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2xkp9" podUID="d4326ced-ae97-4301-854b-222062d4764e" Dec 06 06:55:21 crc kubenswrapper[4945]: I1206 06:55:21.953070 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:21 crc kubenswrapper[4945]: I1206 06:55:21.953132 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:21 crc kubenswrapper[4945]: I1206 06:55:21.953105 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:21 crc kubenswrapper[4945]: E1206 06:55:21.953300 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 06:55:21 crc kubenswrapper[4945]: E1206 06:55:21.953412 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 06:55:21 crc kubenswrapper[4945]: E1206 06:55:21.953574 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 06:55:22 crc kubenswrapper[4945]: I1206 06:55:22.952914 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:22 crc kubenswrapper[4945]: I1206 06:55:22.955344 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 06:55:22 crc kubenswrapper[4945]: I1206 06:55:22.955603 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 06:55:23 crc kubenswrapper[4945]: I1206 06:55:23.952792 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:23 crc kubenswrapper[4945]: I1206 06:55:23.952892 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:23 crc kubenswrapper[4945]: I1206 06:55:23.952914 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:23 crc kubenswrapper[4945]: I1206 06:55:23.955436 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 06:55:23 crc kubenswrapper[4945]: I1206 06:55:23.956194 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 06:55:23 crc kubenswrapper[4945]: I1206 06:55:23.956710 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 06:55:23 crc kubenswrapper[4945]: I1206 06:55:23.958072 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 06:55:25 crc kubenswrapper[4945]: I1206 06:55:25.973919 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.035756 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x95jj"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.036197 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.039673 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.040564 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.041496 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4qjdx"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.042397 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.043007 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ktrcx"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.043962 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.044862 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.045486 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.045694 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.045985 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.052193 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.052197 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.055880 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.078129 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.078545 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.078738 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.078811 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.078843 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.078870 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.078908 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.078557 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079018 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079047 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079332 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079350 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079407 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079450 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079608 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079694 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079821 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079832 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.079941 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.080292 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.080409 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.080583 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.080792 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.080943 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.086162 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.086818 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.087310 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.092146 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.092156 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.092390 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.092525 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.092669 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.093254 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.093534 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-456vw"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.093894 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.094032 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.094201 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.094378 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.094463 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.094556 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.094569 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.095626 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.101966 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.101997 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.103153 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.105620 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.107706 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-p8hsr"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.108347 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.110394 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.110738 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.111834 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bs8n9"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.112123 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.112549 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.112592 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.113193 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.113509 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.113630 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.117819 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.118673 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.121777 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.122033 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.122085 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.122260 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.122452 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.122579 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.122767 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.122814 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.122973 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.123001 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.123235 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.123252 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.123462 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.123479 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.124124 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.124357 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.124360 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.124607 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.124740 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.124745 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.125020 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.125142 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.125330 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.125392 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.125480 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.125622 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.125728 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.125903 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.126128 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.126904 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.128222 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.128767 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.129367 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.130109 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.130734 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.135069 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.136160 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.146755 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.148608 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.149995 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.150342 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.151147 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.152526 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.152954 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.153179 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.153788 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4qjdx"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.153817 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-j2tjw"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.154472 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.154953 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.156437 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.156534 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171204 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fd597d15-c58a-4569-9482-70a16d2a3133-encryption-config\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171243 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171293 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4sf8\" (UniqueName: \"kubernetes.io/projected/e06b2844-0600-4df6-bf5e-749ed15e596a-kube-api-access-n4sf8\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171316 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171342 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4e308683-ec1e-400b-8503-e671d8ac4589-images\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171362 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3254c754-c3b0-41c7-93d7-f52776056d30-audit-policies\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171377 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-config\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171393 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-metrics-tls\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171409 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e308683-ec1e-400b-8503-e671d8ac4589-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171427 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171444 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-config\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171462 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3254c754-c3b0-41c7-93d7-f52776056d30-serving-cert\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171478 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q7tx\" (UniqueName: \"kubernetes.io/projected/3254c754-c3b0-41c7-93d7-f52776056d30-kube-api-access-5q7tx\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171558 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd618c39-c17a-40d9-add1-f87a6f646241-serving-cert\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171632 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-image-import-ca\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171665 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkh2m\" (UniqueName: \"kubernetes.io/projected/dd618c39-c17a-40d9-add1-f87a6f646241-kube-api-access-fkh2m\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171682 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd597d15-c58a-4569-9482-70a16d2a3133-serving-cert\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171698 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3254c754-c3b0-41c7-93d7-f52776056d30-encryption-config\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171715 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-audit\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171734 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmxsj\" (UniqueName: \"kubernetes.io/projected/c0b86555-314f-41f4-8a5d-5aeb356fc01f-kube-api-access-lmxsj\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171752 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3254c754-c3b0-41c7-93d7-f52776056d30-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171778 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-725tk\" (UniqueName: \"kubernetes.io/projected/fd597d15-c58a-4569-9482-70a16d2a3133-kube-api-access-725tk\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171804 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-trusted-ca\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171861 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0b86555-314f-41f4-8a5d-5aeb356fc01f-auth-proxy-config\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171889 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171898 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4k95\" (UniqueName: \"kubernetes.io/projected/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-kube-api-access-h4k95\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171916 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171933 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-client-ca\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171951 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-config\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171968 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c0b86555-314f-41f4-8a5d-5aeb356fc01f-machine-approver-tls\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.171985 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172004 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3254c754-c3b0-41c7-93d7-f52776056d30-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172021 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172037 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172054 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172070 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e06b2844-0600-4df6-bf5e-749ed15e596a-serving-cert\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172089 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b86555-314f-41f4-8a5d-5aeb356fc01f-config\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172106 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172125 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-client-ca\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172142 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fd597d15-c58a-4569-9482-70a16d2a3133-node-pullsecrets\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172160 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-etcd-serving-ca\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172177 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3254c754-c3b0-41c7-93d7-f52776056d30-audit-dir\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172196 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172215 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd597d15-c58a-4569-9482-70a16d2a3133-etcd-client\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172231 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172247 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172262 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-dir\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172302 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzpw6\" (UniqueName: \"kubernetes.io/projected/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-kube-api-access-jzpw6\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172319 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3254c754-c3b0-41c7-93d7-f52776056d30-etcd-client\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172336 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172356 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172370 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-policies\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172388 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fd597d15-c58a-4569-9482-70a16d2a3133-audit-dir\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172404 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e308683-ec1e-400b-8503-e671d8ac4589-config\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.172421 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz4n5\" (UniqueName: \"kubernetes.io/projected/4e308683-ec1e-400b-8503-e671d8ac4589-kube-api-access-nz4n5\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.174228 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-djxqd"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.175067 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.175351 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.175637 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.175747 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.175802 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.175781 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.176351 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.176718 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.176834 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.176932 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.177025 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.177117 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.177206 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.177441 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.177521 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.177699 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.177823 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.178096 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.177539 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.178719 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.179817 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.182160 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.190059 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.190186 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.190800 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.183521 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.191527 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.185726 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.196246 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pfkqs"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.198838 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mgmz8"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.199146 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.200027 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.204812 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.205072 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.205887 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.206076 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.206467 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-lpl6d"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.206819 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8skx"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.207153 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.207223 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.207257 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.207341 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-lpl6d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.207961 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.208107 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.208476 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.208572 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.208998 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dvn5j"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.209170 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.232202 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.233358 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.235218 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.238153 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.238713 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.239342 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5b5kg"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.239628 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.239834 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.240417 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.242869 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8ppq"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.243675 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.244977 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x95jj"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.246660 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rnq6j"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.248087 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.248432 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bs8n9"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.250680 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-456vw"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.250718 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.252419 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.253856 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.257250 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.258964 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.259205 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.261769 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-djxqd"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.261975 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pfkqs"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.265791 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-j2tjw"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.266832 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ktrcx"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.269256 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.270596 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bl6dw"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.271676 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.272676 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273502 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fd597d15-c58a-4569-9482-70a16d2a3133-audit-dir\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273554 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e308683-ec1e-400b-8503-e671d8ac4589-config\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273584 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-policies\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273626 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17dcd89e-7e17-408c-b7d8-806425dafbb9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-sg6tn\" (UID: \"17dcd89e-7e17-408c-b7d8-806425dafbb9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273643 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fd597d15-c58a-4569-9482-70a16d2a3133-audit-dir\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273658 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bwfn\" (UniqueName: \"kubernetes.io/projected/91c9713c-0fd9-41fb-a4dd-b4d5bc691131-kube-api-access-4bwfn\") pod \"cluster-samples-operator-665b6dd947-dhk64\" (UID: \"91c9713c-0fd9-41fb-a4dd-b4d5bc691131\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273687 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-stats-auth\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273713 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5ab548fc-3aaf-4aef-9982-64bb1c243e90-etcd-service-ca\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273740 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pvd7\" (UniqueName: \"kubernetes.io/projected/e86a5672-d1fa-4706-a073-90193128f097-kube-api-access-5pvd7\") pod \"machine-config-controller-84d6567774-nhwkm\" (UID: \"e86a5672-d1fa-4706-a073-90193128f097\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273766 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz4n5\" (UniqueName: \"kubernetes.io/projected/4e308683-ec1e-400b-8503-e671d8ac4589-kube-api-access-nz4n5\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273794 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11106fdc-adfc-4dd8-8011-599912030ba5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-clfbq\" (UID: \"11106fdc-adfc-4dd8-8011-599912030ba5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17dcd89e-7e17-408c-b7d8-806425dafbb9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-sg6tn\" (UID: \"17dcd89e-7e17-408c-b7d8-806425dafbb9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273852 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/23ec3b33-0492-49bd-b1df-aa5aaac01315-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jv77f\" (UID: \"23ec3b33-0492-49bd-b1df-aa5aaac01315\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273886 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fd597d15-c58a-4569-9482-70a16d2a3133-encryption-config\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273910 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273940 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11106fdc-adfc-4dd8-8011-599912030ba5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-clfbq\" (UID: \"11106fdc-adfc-4dd8-8011-599912030ba5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273968 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zm5l\" (UniqueName: \"kubernetes.io/projected/b6b25270-c23f-4cbb-b697-82ead103ae98-kube-api-access-5zm5l\") pod \"catalog-operator-68c6474976-kbjdx\" (UID: \"b6b25270-c23f-4cbb-b697-82ead103ae98\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.273992 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp29f\" (UniqueName: \"kubernetes.io/projected/23ec3b33-0492-49bd-b1df-aa5aaac01315-kube-api-access-cp29f\") pod \"control-plane-machine-set-operator-78cbb6b69f-jv77f\" (UID: \"23ec3b33-0492-49bd-b1df-aa5aaac01315\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274040 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4sf8\" (UniqueName: \"kubernetes.io/projected/e06b2844-0600-4df6-bf5e-749ed15e596a-kube-api-access-n4sf8\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274066 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274102 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fc7fa8b-727b-439a-bf88-7210f1387f6d-serving-cert\") pod \"openshift-config-operator-7777fb866f-zxzv5\" (UID: \"5fc7fa8b-727b-439a-bf88-7210f1387f6d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274130 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scg9h\" (UniqueName: \"kubernetes.io/projected/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-kube-api-access-scg9h\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274159 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4e308683-ec1e-400b-8503-e671d8ac4589-images\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274183 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3254c754-c3b0-41c7-93d7-f52776056d30-audit-policies\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274203 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-config\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274233 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-metrics-tls\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274246 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-policies\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274260 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e308683-ec1e-400b-8503-e671d8ac4589-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274331 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274366 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmfp6\" (UniqueName: \"kubernetes.io/projected/50e16132-af33-4dbf-b4f4-a9b0c8554e24-kube-api-access-rmfp6\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274391 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-config\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274414 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3254c754-c3b0-41c7-93d7-f52776056d30-serving-cert\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274432 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6b25270-c23f-4cbb-b697-82ead103ae98-srv-cert\") pod \"catalog-operator-68c6474976-kbjdx\" (UID: \"b6b25270-c23f-4cbb-b697-82ead103ae98\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274482 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-oauth-serving-cert\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274522 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q7tx\" (UniqueName: \"kubernetes.io/projected/3254c754-c3b0-41c7-93d7-f52776056d30-kube-api-access-5q7tx\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274546 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tmqw\" (UniqueName: \"kubernetes.io/projected/5d3ba584-a688-4fed-bba4-4a1b5e374703-kube-api-access-4tmqw\") pod \"dns-operator-744455d44c-djxqd\" (UID: \"5d3ba584-a688-4fed-bba4-4a1b5e374703\") " pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274567 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-oauth-config\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274591 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-metrics-certs\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274616 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd69d15e-5487-4c83-8f34-175811403e6a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-n2zb6\" (UID: \"dd69d15e-5487-4c83-8f34-175811403e6a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274640 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66de2d9c-8515-4ba8-ae01-100732a4e407-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4c2s9\" (UID: \"66de2d9c-8515-4ba8-ae01-100732a4e407\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274662 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-trusted-ca-bundle\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274682 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd618c39-c17a-40d9-add1-f87a6f646241-serving-cert\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274703 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-image-import-ca\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274727 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdmqn\" (UniqueName: \"kubernetes.io/projected/5fc7fa8b-727b-439a-bf88-7210f1387f6d-kube-api-access-mdmqn\") pod \"openshift-config-operator-7777fb866f-zxzv5\" (UID: \"5fc7fa8b-727b-439a-bf88-7210f1387f6d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274748 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17dcd89e-7e17-408c-b7d8-806425dafbb9-config\") pod \"kube-apiserver-operator-766d6c64bb-sg6tn\" (UID: \"17dcd89e-7e17-408c-b7d8-806425dafbb9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274767 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkh2m\" (UniqueName: \"kubernetes.io/projected/dd618c39-c17a-40d9-add1-f87a6f646241-kube-api-access-fkh2m\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274786 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd597d15-c58a-4569-9482-70a16d2a3133-serving-cert\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274806 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3254c754-c3b0-41c7-93d7-f52776056d30-encryption-config\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274830 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-audit\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274851 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c46fj\" (UniqueName: \"kubernetes.io/projected/df308111-7e73-4d70-b8f5-7c31b6ed3cda-kube-api-access-c46fj\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfp2d\" (UID: \"df308111-7e73-4d70-b8f5-7c31b6ed3cda\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274880 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmxsj\" (UniqueName: \"kubernetes.io/projected/c0b86555-314f-41f4-8a5d-5aeb356fc01f-kube-api-access-lmxsj\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274898 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3254c754-c3b0-41c7-93d7-f52776056d30-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274917 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ab548fc-3aaf-4aef-9982-64bb1c243e90-serving-cert\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274935 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj6gg\" (UniqueName: \"kubernetes.io/projected/5ab548fc-3aaf-4aef-9982-64bb1c243e90-kube-api-access-cj6gg\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274969 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-images\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.274991 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66de2d9c-8515-4ba8-ae01-100732a4e407-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4c2s9\" (UID: \"66de2d9c-8515-4ba8-ae01-100732a4e407\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275012 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-config\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275034 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgvjb\" (UniqueName: \"kubernetes.io/projected/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-kube-api-access-wgvjb\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275062 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-725tk\" (UniqueName: \"kubernetes.io/projected/fd597d15-c58a-4569-9482-70a16d2a3133-kube-api-access-725tk\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275085 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-trusted-ca\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275106 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0b86555-314f-41f4-8a5d-5aeb356fc01f-auth-proxy-config\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275131 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275167 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4k95\" (UniqueName: \"kubernetes.io/projected/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-kube-api-access-h4k95\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275188 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-default-certificate\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275214 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275232 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-proxy-tls\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275249 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-service-ca-bundle\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275273 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-client-ca\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275337 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-config\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275356 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c0b86555-314f-41f4-8a5d-5aeb356fc01f-machine-approver-tls\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275374 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275395 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd69d15e-5487-4c83-8f34-175811403e6a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-n2zb6\" (UID: \"dd69d15e-5487-4c83-8f34-175811403e6a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275419 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7d28b3c8-4c1d-49b4-bc26-47cdb9867e53-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pfkqs\" (UID: \"7d28b3c8-4c1d-49b4-bc26-47cdb9867e53\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275440 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cs9f\" (UniqueName: \"kubernetes.io/projected/7d28b3c8-4c1d-49b4-bc26-47cdb9867e53-kube-api-access-4cs9f\") pod \"multus-admission-controller-857f4d67dd-pfkqs\" (UID: \"7d28b3c8-4c1d-49b4-bc26-47cdb9867e53\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275460 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e86a5672-d1fa-4706-a073-90193128f097-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nhwkm\" (UID: \"e86a5672-d1fa-4706-a073-90193128f097\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275480 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-serving-cert\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275502 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3254c754-c3b0-41c7-93d7-f52776056d30-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275522 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df308111-7e73-4d70-b8f5-7c31b6ed3cda-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfp2d\" (UID: \"df308111-7e73-4d70-b8f5-7c31b6ed3cda\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275544 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275569 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5ab548fc-3aaf-4aef-9982-64bb1c243e90-etcd-ca\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275587 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8hxl\" (UniqueName: \"kubernetes.io/projected/5b5df8d0-7bf5-478b-adb7-eb7291da8935-kube-api-access-m8hxl\") pod \"migrator-59844c95c7-ggwcv\" (UID: \"5b5df8d0-7bf5-478b-adb7-eb7291da8935\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275606 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66de2d9c-8515-4ba8-ae01-100732a4e407-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4c2s9\" (UID: \"66de2d9c-8515-4ba8-ae01-100732a4e407\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275623 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6b25270-c23f-4cbb-b697-82ead103ae98-profile-collector-cert\") pod \"catalog-operator-68c6474976-kbjdx\" (UID: \"b6b25270-c23f-4cbb-b697-82ead103ae98\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275641 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275658 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275678 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5fc7fa8b-727b-439a-bf88-7210f1387f6d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zxzv5\" (UID: \"5fc7fa8b-727b-439a-bf88-7210f1387f6d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275697 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e86a5672-d1fa-4706-a073-90193128f097-proxy-tls\") pod \"machine-config-controller-84d6567774-nhwkm\" (UID: \"e86a5672-d1fa-4706-a073-90193128f097\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275717 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-service-ca\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275737 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e06b2844-0600-4df6-bf5e-749ed15e596a-serving-cert\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275756 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b86555-314f-41f4-8a5d-5aeb356fc01f-config\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275774 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275793 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab548fc-3aaf-4aef-9982-64bb1c243e90-config\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275817 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fd597d15-c58a-4569-9482-70a16d2a3133-node-pullsecrets\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275838 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-etcd-serving-ca\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275860 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3254c754-c3b0-41c7-93d7-f52776056d30-audit-dir\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275885 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5ab548fc-3aaf-4aef-9982-64bb1c243e90-etcd-client\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275903 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5d3ba584-a688-4fed-bba4-4a1b5e374703-metrics-tls\") pod \"dns-operator-744455d44c-djxqd\" (UID: \"5d3ba584-a688-4fed-bba4-4a1b5e374703\") " pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275923 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-client-ca\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275940 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd597d15-c58a-4569-9482-70a16d2a3133-etcd-client\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275962 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.275986 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.276005 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-dir\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.276028 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/91c9713c-0fd9-41fb-a4dd-b4d5bc691131-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dhk64\" (UID: \"91c9713c-0fd9-41fb-a4dd-b4d5bc691131\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.276045 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.276063 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzpw6\" (UniqueName: \"kubernetes.io/projected/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-kube-api-access-jzpw6\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.276082 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df308111-7e73-4d70-b8f5-7c31b6ed3cda-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfp2d\" (UID: \"df308111-7e73-4d70-b8f5-7c31b6ed3cda\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.276112 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.276133 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.276152 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd69d15e-5487-4c83-8f34-175811403e6a-config\") pod \"kube-controller-manager-operator-78b949d7b-n2zb6\" (UID: \"dd69d15e-5487-4c83-8f34-175811403e6a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.276175 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5nsg\" (UniqueName: \"kubernetes.io/projected/11106fdc-adfc-4dd8-8011-599912030ba5-kube-api-access-c5nsg\") pod \"kube-storage-version-migrator-operator-b67b599dd-clfbq\" (UID: \"11106fdc-adfc-4dd8-8011-599912030ba5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.276246 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3254c754-c3b0-41c7-93d7-f52776056d30-etcd-client\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.277470 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e308683-ec1e-400b-8503-e671d8ac4589-config\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.278396 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-image-import-ca\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.280299 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fd597d15-c58a-4569-9482-70a16d2a3133-node-pullsecrets\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.280517 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3254c754-c3b0-41c7-93d7-f52776056d30-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.280992 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.284634 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-client-ca\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.286996 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.287012 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-etcd-serving-ca\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.287501 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e06b2844-0600-4df6-bf5e-749ed15e596a-serving-cert\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.287799 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-config\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.289103 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd597d15-c58a-4569-9482-70a16d2a3133-serving-cert\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.289942 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3254c754-c3b0-41c7-93d7-f52776056d30-audit-dir\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.293700 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b86555-314f-41f4-8a5d-5aeb356fc01f-config\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.290590 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-metrics-tls\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.290739 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-dir\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.290780 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.293796 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8skx"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.292864 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.293331 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.293473 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0b86555-314f-41f4-8a5d-5aeb356fc01f-auth-proxy-config\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.290884 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3254c754-c3b0-41c7-93d7-f52776056d30-audit-policies\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.290002 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.294297 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.294431 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fd597d15-c58a-4569-9482-70a16d2a3133-encryption-config\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.294455 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.294531 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.295406 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.296732 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3254c754-c3b0-41c7-93d7-f52776056d30-etcd-client\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.296864 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e308683-ec1e-400b-8503-e671d8ac4589-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.297363 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3254c754-c3b0-41c7-93d7-f52776056d30-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.297506 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.297714 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4e308683-ec1e-400b-8503-e671d8ac4589-images\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.297799 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-config\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.297859 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.297956 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd618c39-c17a-40d9-add1-f87a6f646241-serving-cert\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.298113 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-audit\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.298523 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd597d15-c58a-4569-9482-70a16d2a3133-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.299035 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.299370 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-config\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.299466 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-client-ca\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.299813 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.300365 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-trusted-ca\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.302655 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-lpl6d"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.303669 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.304083 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c0b86555-314f-41f4-8a5d-5aeb356fc01f-machine-approver-tls\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.305570 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.305988 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3254c754-c3b0-41c7-93d7-f52776056d30-serving-cert\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.306086 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd597d15-c58a-4569-9482-70a16d2a3133-etcd-client\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.310225 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.321149 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.323853 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.337457 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.339374 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.341032 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.341747 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.343439 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3254c754-c3b0-41c7-93d7-f52776056d30-encryption-config\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.344366 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mgmz8"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.345605 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.347216 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rnq6j"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.349340 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.350540 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dvn5j"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.352188 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.352901 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8ppq"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.354166 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.355243 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.357633 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.358687 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5b5kg"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.358906 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.359783 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-kngc2"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.360549 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kngc2" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.360714 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cf4n5"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.362318 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.363497 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kngc2"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.365126 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cf4n5"] Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.378999 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384068 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd69d15e-5487-4c83-8f34-175811403e6a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-n2zb6\" (UID: \"dd69d15e-5487-4c83-8f34-175811403e6a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384097 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7d28b3c8-4c1d-49b4-bc26-47cdb9867e53-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pfkqs\" (UID: \"7d28b3c8-4c1d-49b4-bc26-47cdb9867e53\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384122 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df308111-7e73-4d70-b8f5-7c31b6ed3cda-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfp2d\" (UID: \"df308111-7e73-4d70-b8f5-7c31b6ed3cda\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384143 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66de2d9c-8515-4ba8-ae01-100732a4e407-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4c2s9\" (UID: \"66de2d9c-8515-4ba8-ae01-100732a4e407\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384160 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5fc7fa8b-727b-439a-bf88-7210f1387f6d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zxzv5\" (UID: \"5fc7fa8b-727b-439a-bf88-7210f1387f6d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384178 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-service-ca\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5d3ba584-a688-4fed-bba4-4a1b5e374703-metrics-tls\") pod \"dns-operator-744455d44c-djxqd\" (UID: \"5d3ba584-a688-4fed-bba4-4a1b5e374703\") " pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384228 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df308111-7e73-4d70-b8f5-7c31b6ed3cda-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfp2d\" (UID: \"df308111-7e73-4d70-b8f5-7c31b6ed3cda\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384248 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5nsg\" (UniqueName: \"kubernetes.io/projected/11106fdc-adfc-4dd8-8011-599912030ba5-kube-api-access-c5nsg\") pod \"kube-storage-version-migrator-operator-b67b599dd-clfbq\" (UID: \"11106fdc-adfc-4dd8-8011-599912030ba5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384270 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pvd7\" (UniqueName: \"kubernetes.io/projected/e86a5672-d1fa-4706-a073-90193128f097-kube-api-access-5pvd7\") pod \"machine-config-controller-84d6567774-nhwkm\" (UID: \"e86a5672-d1fa-4706-a073-90193128f097\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384332 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17dcd89e-7e17-408c-b7d8-806425dafbb9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-sg6tn\" (UID: \"17dcd89e-7e17-408c-b7d8-806425dafbb9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384352 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bwfn\" (UniqueName: \"kubernetes.io/projected/91c9713c-0fd9-41fb-a4dd-b4d5bc691131-kube-api-access-4bwfn\") pod \"cluster-samples-operator-665b6dd947-dhk64\" (UID: \"91c9713c-0fd9-41fb-a4dd-b4d5bc691131\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384369 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11106fdc-adfc-4dd8-8011-599912030ba5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-clfbq\" (UID: \"11106fdc-adfc-4dd8-8011-599912030ba5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384388 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/23ec3b33-0492-49bd-b1df-aa5aaac01315-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jv77f\" (UID: \"23ec3b33-0492-49bd-b1df-aa5aaac01315\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384427 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmfp6\" (UniqueName: \"kubernetes.io/projected/50e16132-af33-4dbf-b4f4-a9b0c8554e24-kube-api-access-rmfp6\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384444 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-oauth-serving-cert\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384483 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66de2d9c-8515-4ba8-ae01-100732a4e407-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4c2s9\" (UID: \"66de2d9c-8515-4ba8-ae01-100732a4e407\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384513 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdmqn\" (UniqueName: \"kubernetes.io/projected/5fc7fa8b-727b-439a-bf88-7210f1387f6d-kube-api-access-mdmqn\") pod \"openshift-config-operator-7777fb866f-zxzv5\" (UID: \"5fc7fa8b-727b-439a-bf88-7210f1387f6d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384534 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ab548fc-3aaf-4aef-9982-64bb1c243e90-serving-cert\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384553 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-images\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384575 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-config\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384594 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-service-ca-bundle\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384612 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e86a5672-d1fa-4706-a073-90193128f097-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nhwkm\" (UID: \"e86a5672-d1fa-4706-a073-90193128f097\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384657 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-serving-cert\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384679 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cs9f\" (UniqueName: \"kubernetes.io/projected/7d28b3c8-4c1d-49b4-bc26-47cdb9867e53-kube-api-access-4cs9f\") pod \"multus-admission-controller-857f4d67dd-pfkqs\" (UID: \"7d28b3c8-4c1d-49b4-bc26-47cdb9867e53\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384699 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5ab548fc-3aaf-4aef-9982-64bb1c243e90-etcd-ca\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384717 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8hxl\" (UniqueName: \"kubernetes.io/projected/5b5df8d0-7bf5-478b-adb7-eb7291da8935-kube-api-access-m8hxl\") pod \"migrator-59844c95c7-ggwcv\" (UID: \"5b5df8d0-7bf5-478b-adb7-eb7291da8935\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384738 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6b25270-c23f-4cbb-b697-82ead103ae98-profile-collector-cert\") pod \"catalog-operator-68c6474976-kbjdx\" (UID: \"b6b25270-c23f-4cbb-b697-82ead103ae98\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384762 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7adf125b-8ab8-42b0-9d4d-c38aacbc1a17-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vj29d\" (UID: \"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384789 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e86a5672-d1fa-4706-a073-90193128f097-proxy-tls\") pod \"machine-config-controller-84d6567774-nhwkm\" (UID: \"e86a5672-d1fa-4706-a073-90193128f097\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7adf125b-8ab8-42b0-9d4d-c38aacbc1a17-srv-cert\") pod \"olm-operator-6b444d44fb-vj29d\" (UID: \"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384834 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab548fc-3aaf-4aef-9982-64bb1c243e90-config\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384851 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5ab548fc-3aaf-4aef-9982-64bb1c243e90-etcd-client\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384867 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/91c9713c-0fd9-41fb-a4dd-b4d5bc691131-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dhk64\" (UID: \"91c9713c-0fd9-41fb-a4dd-b4d5bc691131\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384884 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd69d15e-5487-4c83-8f34-175811403e6a-config\") pod \"kube-controller-manager-operator-78b949d7b-n2zb6\" (UID: \"dd69d15e-5487-4c83-8f34-175811403e6a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384901 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5ab548fc-3aaf-4aef-9982-64bb1c243e90-etcd-service-ca\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384916 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-stats-auth\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384937 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17dcd89e-7e17-408c-b7d8-806425dafbb9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-sg6tn\" (UID: \"17dcd89e-7e17-408c-b7d8-806425dafbb9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384952 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11106fdc-adfc-4dd8-8011-599912030ba5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-clfbq\" (UID: \"11106fdc-adfc-4dd8-8011-599912030ba5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384969 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zm5l\" (UniqueName: \"kubernetes.io/projected/b6b25270-c23f-4cbb-b697-82ead103ae98-kube-api-access-5zm5l\") pod \"catalog-operator-68c6474976-kbjdx\" (UID: \"b6b25270-c23f-4cbb-b697-82ead103ae98\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.384990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp29f\" (UniqueName: \"kubernetes.io/projected/23ec3b33-0492-49bd-b1df-aa5aaac01315-kube-api-access-cp29f\") pod \"control-plane-machine-set-operator-78cbb6b69f-jv77f\" (UID: \"23ec3b33-0492-49bd-b1df-aa5aaac01315\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385016 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fc7fa8b-727b-439a-bf88-7210f1387f6d-serving-cert\") pod \"openshift-config-operator-7777fb866f-zxzv5\" (UID: \"5fc7fa8b-727b-439a-bf88-7210f1387f6d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385032 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scg9h\" (UniqueName: \"kubernetes.io/projected/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-kube-api-access-scg9h\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385049 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6b25270-c23f-4cbb-b697-82ead103ae98-srv-cert\") pod \"catalog-operator-68c6474976-kbjdx\" (UID: \"b6b25270-c23f-4cbb-b697-82ead103ae98\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385067 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-oauth-config\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385083 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-metrics-certs\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385099 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tmqw\" (UniqueName: \"kubernetes.io/projected/5d3ba584-a688-4fed-bba4-4a1b5e374703-kube-api-access-4tmqw\") pod \"dns-operator-744455d44c-djxqd\" (UID: \"5d3ba584-a688-4fed-bba4-4a1b5e374703\") " pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385117 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd69d15e-5487-4c83-8f34-175811403e6a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-n2zb6\" (UID: \"dd69d15e-5487-4c83-8f34-175811403e6a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385133 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-trusted-ca-bundle\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385151 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-node-bootstrap-token\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385169 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxrhs\" (UniqueName: \"kubernetes.io/projected/7adf125b-8ab8-42b0-9d4d-c38aacbc1a17-kube-api-access-mxrhs\") pod \"olm-operator-6b444d44fb-vj29d\" (UID: \"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385191 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17dcd89e-7e17-408c-b7d8-806425dafbb9-config\") pod \"kube-apiserver-operator-766d6c64bb-sg6tn\" (UID: \"17dcd89e-7e17-408c-b7d8-806425dafbb9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385206 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-certs\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385222 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjjjg\" (UniqueName: \"kubernetes.io/projected/28449122-a88f-42b7-8d19-c1875c7b329d-kube-api-access-qjjjg\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385238 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c46fj\" (UniqueName: \"kubernetes.io/projected/df308111-7e73-4d70-b8f5-7c31b6ed3cda-kube-api-access-c46fj\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfp2d\" (UID: \"df308111-7e73-4d70-b8f5-7c31b6ed3cda\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385261 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj6gg\" (UniqueName: \"kubernetes.io/projected/5ab548fc-3aaf-4aef-9982-64bb1c243e90-kube-api-access-cj6gg\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385309 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66de2d9c-8515-4ba8-ae01-100732a4e407-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4c2s9\" (UID: \"66de2d9c-8515-4ba8-ae01-100732a4e407\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385333 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgvjb\" (UniqueName: \"kubernetes.io/projected/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-kube-api-access-wgvjb\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385361 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385381 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm5pf\" (UniqueName: \"kubernetes.io/projected/5bfd04a8-aad7-482a-9e68-f74c9cff00ba-kube-api-access-wm5pf\") pod \"downloads-7954f5f757-lpl6d\" (UID: \"5bfd04a8-aad7-482a-9e68-f74c9cff00ba\") " pod="openshift-console/downloads-7954f5f757-lpl6d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385403 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-default-certificate\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.385421 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-proxy-tls\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.386942 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17dcd89e-7e17-408c-b7d8-806425dafbb9-config\") pod \"kube-apiserver-operator-766d6c64bb-sg6tn\" (UID: \"17dcd89e-7e17-408c-b7d8-806425dafbb9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.387032 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e86a5672-d1fa-4706-a073-90193128f097-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nhwkm\" (UID: \"e86a5672-d1fa-4706-a073-90193128f097\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.387781 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.387917 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-trusted-ca-bundle\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.388147 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5ab548fc-3aaf-4aef-9982-64bb1c243e90-etcd-service-ca\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.388206 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd69d15e-5487-4c83-8f34-175811403e6a-config\") pod \"kube-controller-manager-operator-78b949d7b-n2zb6\" (UID: \"dd69d15e-5487-4c83-8f34-175811403e6a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.389304 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-service-ca\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.389391 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5ab548fc-3aaf-4aef-9982-64bb1c243e90-etcd-ca\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.389420 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-service-ca-bundle\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.389541 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-oauth-config\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.389892 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-oauth-serving-cert\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.390096 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd69d15e-5487-4c83-8f34-175811403e6a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-n2zb6\" (UID: \"dd69d15e-5487-4c83-8f34-175811403e6a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.390592 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11106fdc-adfc-4dd8-8011-599912030ba5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-clfbq\" (UID: \"11106fdc-adfc-4dd8-8011-599912030ba5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.391094 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5fc7fa8b-727b-439a-bf88-7210f1387f6d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zxzv5\" (UID: \"5fc7fa8b-727b-439a-bf88-7210f1387f6d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.391297 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab548fc-3aaf-4aef-9982-64bb1c243e90-config\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.391894 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-config\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.392399 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17dcd89e-7e17-408c-b7d8-806425dafbb9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-sg6tn\" (UID: \"17dcd89e-7e17-408c-b7d8-806425dafbb9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.392446 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5ab548fc-3aaf-4aef-9982-64bb1c243e90-etcd-client\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.392584 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-stats-auth\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.392958 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/91c9713c-0fd9-41fb-a4dd-b4d5bc691131-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dhk64\" (UID: \"91c9713c-0fd9-41fb-a4dd-b4d5bc691131\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.394308 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/23ec3b33-0492-49bd-b1df-aa5aaac01315-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jv77f\" (UID: \"23ec3b33-0492-49bd-b1df-aa5aaac01315\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.394379 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-metrics-certs\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.394380 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-serving-cert\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.394461 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e86a5672-d1fa-4706-a073-90193128f097-proxy-tls\") pod \"machine-config-controller-84d6567774-nhwkm\" (UID: \"e86a5672-d1fa-4706-a073-90193128f097\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.395351 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ab548fc-3aaf-4aef-9982-64bb1c243e90-serving-cert\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.395530 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11106fdc-adfc-4dd8-8011-599912030ba5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-clfbq\" (UID: \"11106fdc-adfc-4dd8-8011-599912030ba5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.395936 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-default-certificate\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.399342 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.408864 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fc7fa8b-727b-439a-bf88-7210f1387f6d-serving-cert\") pod \"openshift-config-operator-7777fb866f-zxzv5\" (UID: \"5fc7fa8b-727b-439a-bf88-7210f1387f6d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.419667 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.439630 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.458835 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.479690 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.485089 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66de2d9c-8515-4ba8-ae01-100732a4e407-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4c2s9\" (UID: \"66de2d9c-8515-4ba8-ae01-100732a4e407\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.486946 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7adf125b-8ab8-42b0-9d4d-c38aacbc1a17-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vj29d\" (UID: \"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.487005 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7adf125b-8ab8-42b0-9d4d-c38aacbc1a17-srv-cert\") pod \"olm-operator-6b444d44fb-vj29d\" (UID: \"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.487110 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-node-bootstrap-token\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.487142 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxrhs\" (UniqueName: \"kubernetes.io/projected/7adf125b-8ab8-42b0-9d4d-c38aacbc1a17-kube-api-access-mxrhs\") pod \"olm-operator-6b444d44fb-vj29d\" (UID: \"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.487173 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-certs\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.487210 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjjjg\" (UniqueName: \"kubernetes.io/projected/28449122-a88f-42b7-8d19-c1875c7b329d-kube-api-access-qjjjg\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.487309 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm5pf\" (UniqueName: \"kubernetes.io/projected/5bfd04a8-aad7-482a-9e68-f74c9cff00ba-kube-api-access-wm5pf\") pod \"downloads-7954f5f757-lpl6d\" (UID: \"5bfd04a8-aad7-482a-9e68-f74c9cff00ba\") " pod="openshift-console/downloads-7954f5f757-lpl6d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.498876 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.520603 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.529401 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66de2d9c-8515-4ba8-ae01-100732a4e407-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4c2s9\" (UID: \"66de2d9c-8515-4ba8-ae01-100732a4e407\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.558978 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.579869 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.599343 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.612324 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5d3ba584-a688-4fed-bba4-4a1b5e374703-metrics-tls\") pod \"dns-operator-744455d44c-djxqd\" (UID: \"5d3ba584-a688-4fed-bba4-4a1b5e374703\") " pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.618554 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.639940 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.640917 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-images\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.660415 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.669916 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-proxy-tls\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.679536 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.699146 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.718838 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.739985 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.759523 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.774708 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df308111-7e73-4d70-b8f5-7c31b6ed3cda-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfp2d\" (UID: \"df308111-7e73-4d70-b8f5-7c31b6ed3cda\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.779935 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.800190 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.809941 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df308111-7e73-4d70-b8f5-7c31b6ed3cda-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfp2d\" (UID: \"df308111-7e73-4d70-b8f5-7c31b6ed3cda\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.818986 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.839029 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.853149 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6b25270-c23f-4cbb-b697-82ead103ae98-srv-cert\") pod \"catalog-operator-68c6474976-kbjdx\" (UID: \"b6b25270-c23f-4cbb-b697-82ead103ae98\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.859758 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.870970 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6b25270-c23f-4cbb-b697-82ead103ae98-profile-collector-cert\") pod \"catalog-operator-68c6474976-kbjdx\" (UID: \"b6b25270-c23f-4cbb-b697-82ead103ae98\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.871910 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7adf125b-8ab8-42b0-9d4d-c38aacbc1a17-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vj29d\" (UID: \"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.879730 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.899467 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.903871 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7d28b3c8-4c1d-49b4-bc26-47cdb9867e53-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pfkqs\" (UID: \"7d28b3c8-4c1d-49b4-bc26-47cdb9867e53\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.920248 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.940177 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.959614 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 06:55:26 crc kubenswrapper[4945]: I1206 06:55:26.979389 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.002053 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.030627 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.039240 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.059243 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.080790 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.098871 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.118678 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.139683 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.159349 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.179444 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.205680 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.217826 4945 request.go:700] Waited for 1.009476693s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/configmaps?fieldSelector=metadata.name%3Dtrusted-ca&limit=500&resourceVersion=0 Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.228119 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.239715 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.258958 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.280263 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.291621 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7adf125b-8ab8-42b0-9d4d-c38aacbc1a17-srv-cert\") pod \"olm-operator-6b444d44fb-vj29d\" (UID: \"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.298832 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.325898 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.343383 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.364957 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.379813 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.399217 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.419008 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.439130 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.458789 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.479546 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 06:55:27 crc kubenswrapper[4945]: E1206 06:55:27.488247 4945 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Dec 06 06:55:27 crc kubenswrapper[4945]: E1206 06:55:27.488306 4945 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Dec 06 06:55:27 crc kubenswrapper[4945]: E1206 06:55:27.488388 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-node-bootstrap-token podName:28449122-a88f-42b7-8d19-c1875c7b329d nodeName:}" failed. No retries permitted until 2025-12-06 06:55:27.988355552 +0000 UTC m=+141.443216606 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-node-bootstrap-token") pod "machine-config-server-bl6dw" (UID: "28449122-a88f-42b7-8d19-c1875c7b329d") : failed to sync secret cache: timed out waiting for the condition Dec 06 06:55:27 crc kubenswrapper[4945]: E1206 06:55:27.488428 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-certs podName:28449122-a88f-42b7-8d19-c1875c7b329d nodeName:}" failed. No retries permitted until 2025-12-06 06:55:27.988411094 +0000 UTC m=+141.443272408 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-certs") pod "machine-config-server-bl6dw" (UID: "28449122-a88f-42b7-8d19-c1875c7b329d") : failed to sync secret cache: timed out waiting for the condition Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.498162 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.519079 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.539937 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.559548 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.579750 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.598902 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.618659 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.639211 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.659012 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.679846 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.699260 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.723360 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.738348 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.759078 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.780359 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.799180 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.818603 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.839813 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.859668 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.879596 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.899206 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.919353 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.939707 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.978186 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz4n5\" (UniqueName: \"kubernetes.io/projected/4e308683-ec1e-400b-8503-e671d8ac4589-kube-api-access-nz4n5\") pod \"machine-api-operator-5694c8668f-4qjdx\" (UID: \"4e308683-ec1e-400b-8503-e671d8ac4589\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:27 crc kubenswrapper[4945]: I1206 06:55:27.996465 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkh2m\" (UniqueName: \"kubernetes.io/projected/dd618c39-c17a-40d9-add1-f87a6f646241-kube-api-access-fkh2m\") pod \"route-controller-manager-6576b87f9c-xrvfp\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.011053 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-node-bootstrap-token\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.011109 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-certs\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.016121 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4k95\" (UniqueName: \"kubernetes.io/projected/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-kube-api-access-h4k95\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.016880 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-certs\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.017849 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/28449122-a88f-42b7-8d19-c1875c7b329d-node-bootstrap-token\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.050070 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-725tk\" (UniqueName: \"kubernetes.io/projected/fd597d15-c58a-4569-9482-70a16d2a3133-kube-api-access-725tk\") pod \"apiserver-76f77b778f-ktrcx\" (UID: \"fd597d15-c58a-4569-9482-70a16d2a3133\") " pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.052135 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q7tx\" (UniqueName: \"kubernetes.io/projected/3254c754-c3b0-41c7-93d7-f52776056d30-kube-api-access-5q7tx\") pod \"apiserver-7bbb656c7d-d2p6r\" (UID: \"3254c754-c3b0-41c7-93d7-f52776056d30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.075425 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmxsj\" (UniqueName: \"kubernetes.io/projected/c0b86555-314f-41f4-8a5d-5aeb356fc01f-kube-api-access-lmxsj\") pod \"machine-approver-56656f9798-gc6qb\" (UID: \"c0b86555-314f-41f4-8a5d-5aeb356fc01f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.092691 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4fdae04d-27bb-4419-88cf-0ff25eeb6ff8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4blg9\" (UID: \"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.114333 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4sf8\" (UniqueName: \"kubernetes.io/projected/e06b2844-0600-4df6-bf5e-749ed15e596a-kube-api-access-n4sf8\") pod \"controller-manager-879f6c89f-x95jj\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.131491 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzpw6\" (UniqueName: \"kubernetes.io/projected/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-kube-api-access-jzpw6\") pod \"oauth-openshift-558db77b4-456vw\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.159368 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.167536 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.179091 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.187570 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.199296 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.204436 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.220125 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.220469 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.237024 4945 request.go:700] Waited for 1.874438353s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.239161 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.244674 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.252630 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.259997 4945 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 06:55:28 crc kubenswrapper[4945]: W1206 06:55:28.274555 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0b86555_314f_41f4_8a5d_5aeb356fc01f.slice/crio-29783c1b9d9b890ca9bf7c3d737a381dcde429df32a898de42baf064fe41ba29 WatchSource:0}: Error finding container 29783c1b9d9b890ca9bf7c3d737a381dcde429df32a898de42baf064fe41ba29: Status 404 returned error can't find the container with id 29783c1b9d9b890ca9bf7c3d737a381dcde429df32a898de42baf064fe41ba29 Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.278406 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.278949 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.288727 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.323632 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c46fj\" (UniqueName: \"kubernetes.io/projected/df308111-7e73-4d70-b8f5-7c31b6ed3cda-kube-api-access-c46fj\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfp2d\" (UID: \"df308111-7e73-4d70-b8f5-7c31b6ed3cda\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.344618 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj6gg\" (UniqueName: \"kubernetes.io/projected/5ab548fc-3aaf-4aef-9982-64bb1c243e90-kube-api-access-cj6gg\") pod \"etcd-operator-b45778765-bs8n9\" (UID: \"5ab548fc-3aaf-4aef-9982-64bb1c243e90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.370597 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp29f\" (UniqueName: \"kubernetes.io/projected/23ec3b33-0492-49bd-b1df-aa5aaac01315-kube-api-access-cp29f\") pod \"control-plane-machine-set-operator-78cbb6b69f-jv77f\" (UID: \"23ec3b33-0492-49bd-b1df-aa5aaac01315\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.375781 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cs9f\" (UniqueName: \"kubernetes.io/projected/7d28b3c8-4c1d-49b4-bc26-47cdb9867e53-kube-api-access-4cs9f\") pod \"multus-admission-controller-857f4d67dd-pfkqs\" (UID: \"7d28b3c8-4c1d-49b4-bc26-47cdb9867e53\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.411113 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66de2d9c-8515-4ba8-ae01-100732a4e407-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4c2s9\" (UID: \"66de2d9c-8515-4ba8-ae01-100732a4e407\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.423605 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd69d15e-5487-4c83-8f34-175811403e6a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-n2zb6\" (UID: \"dd69d15e-5487-4c83-8f34-175811403e6a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.435210 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8hxl\" (UniqueName: \"kubernetes.io/projected/5b5df8d0-7bf5-478b-adb7-eb7291da8935-kube-api-access-m8hxl\") pod \"migrator-59844c95c7-ggwcv\" (UID: \"5b5df8d0-7bf5-478b-adb7-eb7291da8935\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.454499 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tmqw\" (UniqueName: \"kubernetes.io/projected/5d3ba584-a688-4fed-bba4-4a1b5e374703-kube-api-access-4tmqw\") pod \"dns-operator-744455d44c-djxqd\" (UID: \"5d3ba584-a688-4fed-bba4-4a1b5e374703\") " pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.468407 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.476968 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.479061 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17dcd89e-7e17-408c-b7d8-806425dafbb9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-sg6tn\" (UID: \"17dcd89e-7e17-408c-b7d8-806425dafbb9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.494196 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.499594 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdmqn\" (UniqueName: \"kubernetes.io/projected/5fc7fa8b-727b-439a-bf88-7210f1387f6d-kube-api-access-mdmqn\") pod \"openshift-config-operator-7777fb866f-zxzv5\" (UID: \"5fc7fa8b-727b-439a-bf88-7210f1387f6d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.510864 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.521600 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bwfn\" (UniqueName: \"kubernetes.io/projected/91c9713c-0fd9-41fb-a4dd-b4d5bc691131-kube-api-access-4bwfn\") pod \"cluster-samples-operator-665b6dd947-dhk64\" (UID: \"91c9713c-0fd9-41fb-a4dd-b4d5bc691131\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.549264 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5nsg\" (UniqueName: \"kubernetes.io/projected/11106fdc-adfc-4dd8-8011-599912030ba5-kube-api-access-c5nsg\") pod \"kube-storage-version-migrator-operator-b67b599dd-clfbq\" (UID: \"11106fdc-adfc-4dd8-8011-599912030ba5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.558997 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgvjb\" (UniqueName: \"kubernetes.io/projected/3e69fe71-4b05-43dd-a6a8-dd10ec56c938-kube-api-access-wgvjb\") pod \"router-default-5444994796-p8hsr\" (UID: \"3e69fe71-4b05-43dd-a6a8-dd10ec56c938\") " pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.587457 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmfp6\" (UniqueName: \"kubernetes.io/projected/50e16132-af33-4dbf-b4f4-a9b0c8554e24-kube-api-access-rmfp6\") pod \"console-f9d7485db-j2tjw\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.599338 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scg9h\" (UniqueName: \"kubernetes.io/projected/2d9fa11b-0c16-4915-b12c-8cc917d82fb7-kube-api-access-scg9h\") pod \"machine-config-operator-74547568cd-d5rzf\" (UID: \"2d9fa11b-0c16-4915-b12c-8cc917d82fb7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.599669 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.606171 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.619271 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pvd7\" (UniqueName: \"kubernetes.io/projected/e86a5672-d1fa-4706-a073-90193128f097-kube-api-access-5pvd7\") pod \"machine-config-controller-84d6567774-nhwkm\" (UID: \"e86a5672-d1fa-4706-a073-90193128f097\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.634999 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zm5l\" (UniqueName: \"kubernetes.io/projected/b6b25270-c23f-4cbb-b697-82ead103ae98-kube-api-access-5zm5l\") pod \"catalog-operator-68c6474976-kbjdx\" (UID: \"b6b25270-c23f-4cbb-b697-82ead103ae98\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.637924 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" Dec 06 06:55:28 crc kubenswrapper[4945]: W1206 06:55:28.638634 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e69fe71_4b05_43dd_a6a8_dd10ec56c938.slice/crio-db49ca9320ee157aa7107ce385823d70d5781f515c4348cf6d9109e7400afddc WatchSource:0}: Error finding container db49ca9320ee157aa7107ce385823d70d5781f515c4348cf6d9109e7400afddc: Status 404 returned error can't find the container with id db49ca9320ee157aa7107ce385823d70d5781f515c4348cf6d9109e7400afddc Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.652930 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.660120 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.663309 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxrhs\" (UniqueName: \"kubernetes.io/projected/7adf125b-8ab8-42b0-9d4d-c38aacbc1a17-kube-api-access-mxrhs\") pod \"olm-operator-6b444d44fb-vj29d\" (UID: \"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.669923 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.676227 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjjjg\" (UniqueName: \"kubernetes.io/projected/28449122-a88f-42b7-8d19-c1875c7b329d-kube-api-access-qjjjg\") pod \"machine-config-server-bl6dw\" (UID: \"28449122-a88f-42b7-8d19-c1875c7b329d\") " pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.697138 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm5pf\" (UniqueName: \"kubernetes.io/projected/5bfd04a8-aad7-482a-9e68-f74c9cff00ba-kube-api-access-wm5pf\") pod \"downloads-7954f5f757-lpl6d\" (UID: \"5bfd04a8-aad7-482a-9e68-f74c9cff00ba\") " pod="openshift-console/downloads-7954f5f757-lpl6d" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.698497 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.708133 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" event={"ID":"c0b86555-314f-41f4-8a5d-5aeb356fc01f","Type":"ContainerStarted","Data":"29783c1b9d9b890ca9bf7c3d737a381dcde429df32a898de42baf064fe41ba29"} Dec 06 06:55:28 crc kubenswrapper[4945]: I1206 06:55:28.709903 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p8hsr" event={"ID":"3e69fe71-4b05-43dd-a6a8-dd10ec56c938","Type":"ContainerStarted","Data":"db49ca9320ee157aa7107ce385823d70d5781f515c4348cf6d9109e7400afddc"} Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.046437 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.046508 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.046699 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.046818 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-lpl6d" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.046909 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bl6dw" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.046730 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.047018 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.046929 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.047108 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.049344 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-registry-tls\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.049435 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/401f27c7-5383-4ade-839b-59a92b137752-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.051375 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcmnq\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-kube-api-access-pcmnq\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.053318 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/401f27c7-5383-4ade-839b-59a92b137752-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.054990 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-bound-sa-token\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.055537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-trusted-ca\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.055760 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.056017 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-registry-certificates\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.061476 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:29.561451205 +0000 UTC m=+143.016312289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.113008 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x95jj"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.124590 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-456vw"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.124685 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.124914 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ktrcx"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.142110 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4qjdx"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.152686 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.156872 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.157739 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.157886 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:29.65786687 +0000 UTC m=+143.112727914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158482 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-bound-sa-token\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158509 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlx68\" (UniqueName: \"kubernetes.io/projected/77e91d35-8908-44ee-b5b8-86cdb9b46726-kube-api-access-zlx68\") pod \"ingress-canary-kngc2\" (UID: \"77e91d35-8908-44ee-b5b8-86cdb9b46726\") " pod="openshift-ingress-canary/ingress-canary-kngc2" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158527 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-595mv\" (UniqueName: \"kubernetes.io/projected/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-kube-api-access-595mv\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158543 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkfc5\" (UniqueName: \"kubernetes.io/projected/9a24d873-f77d-43d9-a402-b7a3d9194496-kube-api-access-bkfc5\") pod \"collect-profiles-29416725-r5wbf\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158564 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf62b\" (UniqueName: \"kubernetes.io/projected/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-kube-api-access-cf62b\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158580 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b79525cf-ee6a-4a3b-a83d-4df325832a72-config\") pod \"service-ca-operator-777779d784-4wqcz\" (UID: \"b79525cf-ee6a-4a3b-a83d-4df325832a72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158605 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a24d873-f77d-43d9-a402-b7a3d9194496-secret-volume\") pod \"collect-profiles-29416725-r5wbf\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158622 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-trusted-ca\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158639 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8194654b-ee85-4388-bc86-633437991bab-config\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158653 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-tmpfs\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158667 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b79525cf-ee6a-4a3b-a83d-4df325832a72-serving-cert\") pod \"service-ca-operator-777779d784-4wqcz\" (UID: \"b79525cf-ee6a-4a3b-a83d-4df325832a72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158682 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r8skx\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158698 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb5dc394-7097-4efd-893a-5869e136d910-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158723 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8194654b-ee85-4388-bc86-633437991bab-serving-cert\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158737 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k4wt\" (UniqueName: \"kubernetes.io/projected/8d7466cb-add7-436c-8641-0bf1bda1afbd-kube-api-access-4k4wt\") pod \"marketplace-operator-79b997595-r8skx\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158761 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158776 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-plugins-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158801 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qsq9\" (UniqueName: \"kubernetes.io/projected/c9ef2f41-8ec0-48fe-8b43-a2d230b60392-kube-api-access-4qsq9\") pod \"package-server-manager-789f6589d5-62dkz\" (UID: \"c9ef2f41-8ec0-48fe-8b43-a2d230b60392\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66j6z\" (UniqueName: \"kubernetes.io/projected/a9ad8684-da2b-4ac8-ab89-a6eb3ab47311-kube-api-access-66j6z\") pod \"openshift-apiserver-operator-796bbdcf4f-5g55f\" (UID: \"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158834 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-csi-data-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158854 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-registry-certificates\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158873 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a24d873-f77d-43d9-a402-b7a3d9194496-config-volume\") pod \"collect-profiles-29416725-r5wbf\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158900 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q2ss\" (UniqueName: \"kubernetes.io/projected/8194654b-ee85-4388-bc86-633437991bab-kube-api-access-9q2ss\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158919 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9ad8684-da2b-4ac8-ab89-a6eb3ab47311-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5g55f\" (UID: \"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158958 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-registry-tls\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158974 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-socket-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.158990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/401f27c7-5383-4ade-839b-59a92b137752-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159008 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w587f\" (UniqueName: \"kubernetes.io/projected/6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c-kube-api-access-w587f\") pod \"dns-default-rnq6j\" (UID: \"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c\") " pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-apiservice-cert\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159067 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw4dm\" (UniqueName: \"kubernetes.io/projected/cb5dc394-7097-4efd-893a-5869e136d910-kube-api-access-nw4dm\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159109 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb5dc394-7097-4efd-893a-5869e136d910-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159126 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcmnq\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-kube-api-access-pcmnq\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159150 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8708c768-ebe0-4070-b7b6-249649beb4f7-config\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159165 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a7045282-9b6c-4821-aa59-d0f45f65682b-signing-cabundle\") pod \"service-ca-9c57cc56f-dvn5j\" (UID: \"a7045282-9b6c-4821-aa59-d0f45f65682b\") " pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159255 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8708c768-ebe0-4070-b7b6-249649beb4f7-trusted-ca\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159272 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c-config-volume\") pod \"dns-default-rnq6j\" (UID: \"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c\") " pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159312 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t79fq\" (UniqueName: \"kubernetes.io/projected/b79525cf-ee6a-4a3b-a83d-4df325832a72-kube-api-access-t79fq\") pod \"service-ca-operator-777779d784-4wqcz\" (UID: \"b79525cf-ee6a-4a3b-a83d-4df325832a72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159344 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a7045282-9b6c-4821-aa59-d0f45f65682b-signing-key\") pod \"service-ca-9c57cc56f-dvn5j\" (UID: \"a7045282-9b6c-4821-aa59-d0f45f65682b\") " pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159362 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qj44\" (UniqueName: \"kubernetes.io/projected/8708c768-ebe0-4070-b7b6-249649beb4f7-kube-api-access-9qj44\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159376 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-webhook-cert\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159405 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/77e91d35-8908-44ee-b5b8-86cdb9b46726-cert\") pod \"ingress-canary-kngc2\" (UID: \"77e91d35-8908-44ee-b5b8-86cdb9b46726\") " pod="openshift-ingress-canary/ingress-canary-kngc2" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r8skx\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159446 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ad8684-da2b-4ac8-ab89-a6eb3ab47311-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5g55f\" (UID: \"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159472 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8194654b-ee85-4388-bc86-633437991bab-service-ca-bundle\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159486 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c-metrics-tls\") pod \"dns-default-rnq6j\" (UID: \"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c\") " pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159503 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/cb5dc394-7097-4efd-893a-5869e136d910-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159518 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-mountpoint-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159535 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/401f27c7-5383-4ade-839b-59a92b137752-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159558 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8194654b-ee85-4388-bc86-633437991bab-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159612 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xgdl\" (UniqueName: \"kubernetes.io/projected/a7045282-9b6c-4821-aa59-d0f45f65682b-kube-api-access-6xgdl\") pod \"service-ca-9c57cc56f-dvn5j\" (UID: \"a7045282-9b6c-4821-aa59-d0f45f65682b\") " pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159628 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8708c768-ebe0-4070-b7b6-249649beb4f7-serving-cert\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159642 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-registration-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.159658 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c9ef2f41-8ec0-48fe-8b43-a2d230b60392-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-62dkz\" (UID: \"c9ef2f41-8ec0-48fe-8b43-a2d230b60392\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.161914 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:29.661891915 +0000 UTC m=+143.116753039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.163265 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-trusted-ca\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.169924 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/401f27c7-5383-4ade-839b-59a92b137752-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.175768 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-registry-certificates\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.177396 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/401f27c7-5383-4ade-839b-59a92b137752-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.180268 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.183386 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-registry-tls\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.193554 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-bound-sa-token\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.199211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcmnq\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-kube-api-access-pcmnq\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.207644 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod957f12af_ad21_4c4a_97a2_ab1b824aa5a0.slice/crio-62a2d51b6a9d7d7381fc1894e81e088920deb46b61702c870bc5daafd1d9d359 WatchSource:0}: Error finding container 62a2d51b6a9d7d7381fc1894e81e088920deb46b61702c870bc5daafd1d9d359: Status 404 returned error can't find the container with id 62a2d51b6a9d7d7381fc1894e81e088920deb46b61702c870bc5daafd1d9d359 Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.207903 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode06b2844_0600_4df6_bf5e_749ed15e596a.slice/crio-cb9c43d89b3347de5f75ba68e4a1cf36a3c36bcdfa7f71d728f266f34ff2b233 WatchSource:0}: Error finding container cb9c43d89b3347de5f75ba68e4a1cf36a3c36bcdfa7f71d728f266f34ff2b233: Status 404 returned error can't find the container with id cb9c43d89b3347de5f75ba68e4a1cf36a3c36bcdfa7f71d728f266f34ff2b233 Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.215242 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66de2d9c_8515_4ba8_ae01_100732a4e407.slice/crio-15a73937653ecd21f0b4930f89070e021ee7a2f306d60666388239941d3ccaec WatchSource:0}: Error finding container 15a73937653ecd21f0b4930f89070e021ee7a2f306d60666388239941d3ccaec: Status 404 returned error can't find the container with id 15a73937653ecd21f0b4930f89070e021ee7a2f306d60666388239941d3ccaec Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261520 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261758 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w587f\" (UniqueName: \"kubernetes.io/projected/6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c-kube-api-access-w587f\") pod \"dns-default-rnq6j\" (UID: \"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c\") " pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261784 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-apiservice-cert\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261807 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-socket-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261833 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw4dm\" (UniqueName: \"kubernetes.io/projected/cb5dc394-7097-4efd-893a-5869e136d910-kube-api-access-nw4dm\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261865 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb5dc394-7097-4efd-893a-5869e136d910-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261887 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8708c768-ebe0-4070-b7b6-249649beb4f7-config\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261907 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a7045282-9b6c-4821-aa59-d0f45f65682b-signing-cabundle\") pod \"service-ca-9c57cc56f-dvn5j\" (UID: \"a7045282-9b6c-4821-aa59-d0f45f65682b\") " pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261939 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8708c768-ebe0-4070-b7b6-249649beb4f7-trusted-ca\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261961 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t79fq\" (UniqueName: \"kubernetes.io/projected/b79525cf-ee6a-4a3b-a83d-4df325832a72-kube-api-access-t79fq\") pod \"service-ca-operator-777779d784-4wqcz\" (UID: \"b79525cf-ee6a-4a3b-a83d-4df325832a72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.261981 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c-config-volume\") pod \"dns-default-rnq6j\" (UID: \"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c\") " pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262010 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a7045282-9b6c-4821-aa59-d0f45f65682b-signing-key\") pod \"service-ca-9c57cc56f-dvn5j\" (UID: \"a7045282-9b6c-4821-aa59-d0f45f65682b\") " pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262041 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qj44\" (UniqueName: \"kubernetes.io/projected/8708c768-ebe0-4070-b7b6-249649beb4f7-kube-api-access-9qj44\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262059 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-webhook-cert\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262080 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/77e91d35-8908-44ee-b5b8-86cdb9b46726-cert\") pod \"ingress-canary-kngc2\" (UID: \"77e91d35-8908-44ee-b5b8-86cdb9b46726\") " pod="openshift-ingress-canary/ingress-canary-kngc2" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262104 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r8skx\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262128 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ad8684-da2b-4ac8-ab89-a6eb3ab47311-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5g55f\" (UID: \"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8194654b-ee85-4388-bc86-633437991bab-service-ca-bundle\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262174 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-mountpoint-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262194 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c-metrics-tls\") pod \"dns-default-rnq6j\" (UID: \"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c\") " pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262215 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/cb5dc394-7097-4efd-893a-5869e136d910-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262240 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8194654b-ee85-4388-bc86-633437991bab-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262303 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xgdl\" (UniqueName: \"kubernetes.io/projected/a7045282-9b6c-4821-aa59-d0f45f65682b-kube-api-access-6xgdl\") pod \"service-ca-9c57cc56f-dvn5j\" (UID: \"a7045282-9b6c-4821-aa59-d0f45f65682b\") " pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262333 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8708c768-ebe0-4070-b7b6-249649beb4f7-serving-cert\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262356 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-registration-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262380 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c9ef2f41-8ec0-48fe-8b43-a2d230b60392-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-62dkz\" (UID: \"c9ef2f41-8ec0-48fe-8b43-a2d230b60392\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262404 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlx68\" (UniqueName: \"kubernetes.io/projected/77e91d35-8908-44ee-b5b8-86cdb9b46726-kube-api-access-zlx68\") pod \"ingress-canary-kngc2\" (UID: \"77e91d35-8908-44ee-b5b8-86cdb9b46726\") " pod="openshift-ingress-canary/ingress-canary-kngc2" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262423 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-595mv\" (UniqueName: \"kubernetes.io/projected/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-kube-api-access-595mv\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262448 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkfc5\" (UniqueName: \"kubernetes.io/projected/9a24d873-f77d-43d9-a402-b7a3d9194496-kube-api-access-bkfc5\") pod \"collect-profiles-29416725-r5wbf\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262470 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf62b\" (UniqueName: \"kubernetes.io/projected/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-kube-api-access-cf62b\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262489 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b79525cf-ee6a-4a3b-a83d-4df325832a72-config\") pod \"service-ca-operator-777779d784-4wqcz\" (UID: \"b79525cf-ee6a-4a3b-a83d-4df325832a72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262519 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a24d873-f77d-43d9-a402-b7a3d9194496-secret-volume\") pod \"collect-profiles-29416725-r5wbf\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262541 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8194654b-ee85-4388-bc86-633437991bab-config\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262561 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-tmpfs\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262581 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b79525cf-ee6a-4a3b-a83d-4df325832a72-serving-cert\") pod \"service-ca-operator-777779d784-4wqcz\" (UID: \"b79525cf-ee6a-4a3b-a83d-4df325832a72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r8skx\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262620 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8194654b-ee85-4388-bc86-633437991bab-serving-cert\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262639 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k4wt\" (UniqueName: \"kubernetes.io/projected/8d7466cb-add7-436c-8641-0bf1bda1afbd-kube-api-access-4k4wt\") pod \"marketplace-operator-79b997595-r8skx\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262661 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb5dc394-7097-4efd-893a-5869e136d910-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262691 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-plugins-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262713 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qsq9\" (UniqueName: \"kubernetes.io/projected/c9ef2f41-8ec0-48fe-8b43-a2d230b60392-kube-api-access-4qsq9\") pod \"package-server-manager-789f6589d5-62dkz\" (UID: \"c9ef2f41-8ec0-48fe-8b43-a2d230b60392\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262735 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66j6z\" (UniqueName: \"kubernetes.io/projected/a9ad8684-da2b-4ac8-ab89-a6eb3ab47311-kube-api-access-66j6z\") pod \"openshift-apiserver-operator-796bbdcf4f-5g55f\" (UID: \"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262754 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-csi-data-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262777 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a24d873-f77d-43d9-a402-b7a3d9194496-config-volume\") pod \"collect-profiles-29416725-r5wbf\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262800 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q2ss\" (UniqueName: \"kubernetes.io/projected/8194654b-ee85-4388-bc86-633437991bab-kube-api-access-9q2ss\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.262823 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9ad8684-da2b-4ac8-ab89-a6eb3ab47311-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5g55f\" (UID: \"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.263544 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:29.763501836 +0000 UTC m=+143.218362940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.264083 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-tmpfs\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.265162 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-plugins-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.268906 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb5dc394-7097-4efd-893a-5869e136d910-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.270494 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8708c768-ebe0-4070-b7b6-249649beb4f7-serving-cert\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.270605 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-registration-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.271163 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a7045282-9b6c-4821-aa59-d0f45f65682b-signing-cabundle\") pod \"service-ca-9c57cc56f-dvn5j\" (UID: \"a7045282-9b6c-4821-aa59-d0f45f65682b\") " pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.272056 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8708c768-ebe0-4070-b7b6-249649beb4f7-trusted-ca\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.273600 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b79525cf-ee6a-4a3b-a83d-4df325832a72-serving-cert\") pod \"service-ca-operator-777779d784-4wqcz\" (UID: \"b79525cf-ee6a-4a3b-a83d-4df325832a72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.273639 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r8skx\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.265297 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-socket-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.274504 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8194654b-ee85-4388-bc86-633437991bab-config\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.274605 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a7045282-9b6c-4821-aa59-d0f45f65682b-signing-key\") pod \"service-ca-9c57cc56f-dvn5j\" (UID: \"a7045282-9b6c-4821-aa59-d0f45f65682b\") " pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.274548 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c9ef2f41-8ec0-48fe-8b43-a2d230b60392-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-62dkz\" (UID: \"c9ef2f41-8ec0-48fe-8b43-a2d230b60392\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.274819 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-csi-data-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.275058 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-mountpoint-dir\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.275184 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b79525cf-ee6a-4a3b-a83d-4df325832a72-config\") pod \"service-ca-operator-777779d784-4wqcz\" (UID: \"b79525cf-ee6a-4a3b-a83d-4df325832a72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.277114 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c-config-volume\") pod \"dns-default-rnq6j\" (UID: \"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c\") " pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.277348 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8194654b-ee85-4388-bc86-633437991bab-serving-cert\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.278169 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/77e91d35-8908-44ee-b5b8-86cdb9b46726-cert\") pod \"ingress-canary-kngc2\" (UID: \"77e91d35-8908-44ee-b5b8-86cdb9b46726\") " pod="openshift-ingress-canary/ingress-canary-kngc2" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.278850 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8194654b-ee85-4388-bc86-633437991bab-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.279477 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c-metrics-tls\") pod \"dns-default-rnq6j\" (UID: \"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c\") " pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.280268 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r8skx\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.298077 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/cb5dc394-7097-4efd-893a-5869e136d910-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.309747 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xgdl\" (UniqueName: \"kubernetes.io/projected/a7045282-9b6c-4821-aa59-d0f45f65682b-kube-api-access-6xgdl\") pod \"service-ca-9c57cc56f-dvn5j\" (UID: \"a7045282-9b6c-4821-aa59-d0f45f65682b\") " pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.319190 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w587f\" (UniqueName: \"kubernetes.io/projected/6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c-kube-api-access-w587f\") pod \"dns-default-rnq6j\" (UID: \"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c\") " pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.336355 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qj44\" (UniqueName: \"kubernetes.io/projected/8708c768-ebe0-4070-b7b6-249649beb4f7-kube-api-access-9qj44\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.358348 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw4dm\" (UniqueName: \"kubernetes.io/projected/cb5dc394-7097-4efd-893a-5869e136d910-kube-api-access-nw4dm\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.364418 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.364893 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:29.864871781 +0000 UTC m=+143.319732825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.374831 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t79fq\" (UniqueName: \"kubernetes.io/projected/b79525cf-ee6a-4a3b-a83d-4df325832a72-kube-api-access-t79fq\") pod \"service-ca-operator-777779d784-4wqcz\" (UID: \"b79525cf-ee6a-4a3b-a83d-4df325832a72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.416930 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf62b\" (UniqueName: \"kubernetes.io/projected/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-kube-api-access-cf62b\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.434005 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-595mv\" (UniqueName: \"kubernetes.io/projected/e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88-kube-api-access-595mv\") pod \"csi-hostpathplugin-cf4n5\" (UID: \"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88\") " pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.465462 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.465854 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:29.965832554 +0000 UTC m=+143.420693598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.470908 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.478663 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.496558 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlx68\" (UniqueName: \"kubernetes.io/projected/77e91d35-8908-44ee-b5b8-86cdb9b46726-kube-api-access-zlx68\") pod \"ingress-canary-kngc2\" (UID: \"77e91d35-8908-44ee-b5b8-86cdb9b46726\") " pod="openshift-ingress-canary/ingress-canary-kngc2" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.505373 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pfkqs"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.508462 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.518742 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.534547 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qsq9\" (UniqueName: \"kubernetes.io/projected/c9ef2f41-8ec0-48fe-8b43-a2d230b60392-kube-api-access-4qsq9\") pod \"package-server-manager-789f6589d5-62dkz\" (UID: \"c9ef2f41-8ec0-48fe-8b43-a2d230b60392\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.537128 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kngc2" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.564651 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.568895 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.569292 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.069262603 +0000 UTC m=+143.524123637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.583253 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-djxqd"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.589353 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.594832 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bs8n9"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.640656 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-lpl6d"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.660194 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9ad8684-da2b-4ac8-ab89-a6eb3ab47311-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5g55f\" (UID: \"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.660374 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8194654b-ee85-4388-bc86-633437991bab-service-ca-bundle\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.660677 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8708c768-ebe0-4070-b7b6-249649beb4f7-config\") pod \"console-operator-58897d9998-mgmz8\" (UID: \"8708c768-ebe0-4070-b7b6-249649beb4f7\") " pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.661310 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a24d873-f77d-43d9-a402-b7a3d9194496-config-volume\") pod \"collect-profiles-29416725-r5wbf\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.661532 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ad8684-da2b-4ac8-ab89-a6eb3ab47311-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5g55f\" (UID: \"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.663742 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a24d873-f77d-43d9-a402-b7a3d9194496-secret-volume\") pod \"collect-profiles-29416725-r5wbf\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.663903 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkfc5\" (UniqueName: \"kubernetes.io/projected/9a24d873-f77d-43d9-a402-b7a3d9194496-kube-api-access-bkfc5\") pod \"collect-profiles-29416725-r5wbf\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.665375 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q2ss\" (UniqueName: \"kubernetes.io/projected/8194654b-ee85-4388-bc86-633437991bab-kube-api-access-9q2ss\") pod \"authentication-operator-69f744f599-5b5kg\" (UID: \"8194654b-ee85-4388-bc86-633437991bab\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.665452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-apiservice-cert\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.669554 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.669817 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.169776045 +0000 UTC m=+143.624637119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.670229 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.670697 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.170684099 +0000 UTC m=+143.625545143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.671082 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9cf8e342-cc9f-4017-87de-ea4f1dcaa396-webhook-cert\") pod \"packageserver-d55dfcdfc-6kqp6\" (UID: \"9cf8e342-cc9f-4017-87de-ea4f1dcaa396\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.671569 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k4wt\" (UniqueName: \"kubernetes.io/projected/8d7466cb-add7-436c-8641-0bf1bda1afbd-kube-api-access-4k4wt\") pod \"marketplace-operator-79b997595-r8skx\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.671884 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.673633 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66j6z\" (UniqueName: \"kubernetes.io/projected/a9ad8684-da2b-4ac8-ab89-a6eb3ab47311-kube-api-access-66j6z\") pod \"openshift-apiserver-operator-796bbdcf4f-5g55f\" (UID: \"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.692504 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb5dc394-7097-4efd-893a-5869e136d910-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2z8p6\" (UID: \"cb5dc394-7097-4efd-893a-5869e136d910\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.697771 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd618c39_c17a_40d9_add1_f87a6f646241.slice/crio-53266cde319546d78ff1e41e87d82bd34f57bed41a8c25dfdee6b6a15ede6da7 WatchSource:0}: Error finding container 53266cde319546d78ff1e41e87d82bd34f57bed41a8c25dfdee6b6a15ede6da7: Status 404 returned error can't find the container with id 53266cde319546d78ff1e41e87d82bd34f57bed41a8c25dfdee6b6a15ede6da7 Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.705976 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fdae04d_27bb_4419_88cf_0ff25eeb6ff8.slice/crio-311113ba131fdb8a857b5341387867b26b591f5170946899d94dc63e38a03fac WatchSource:0}: Error finding container 311113ba131fdb8a857b5341387867b26b591f5170946899d94dc63e38a03fac: Status 404 returned error can't find the container with id 311113ba131fdb8a857b5341387867b26b591f5170946899d94dc63e38a03fac Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.712927 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3254c754_c3b0_41c7_93d7_f52776056d30.slice/crio-35fb73f0e10834299201127e329ad805b6d4e6654c6ac474c30a8acf91c54638 WatchSource:0}: Error finding container 35fb73f0e10834299201127e329ad805b6d4e6654c6ac474c30a8acf91c54638: Status 404 returned error can't find the container with id 35fb73f0e10834299201127e329ad805b6d4e6654c6ac474c30a8acf91c54638 Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.715670 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d28b3c8_4c1d_49b4_bc26_47cdb9867e53.slice/crio-cd7e583cc4cebbf82abbf9b36862237ea0ef970397d471b472980281b67dcb7b WatchSource:0}: Error finding container cd7e583cc4cebbf82abbf9b36862237ea0ef970397d471b472980281b67dcb7b: Status 404 returned error can't find the container with id cd7e583cc4cebbf82abbf9b36862237ea0ef970397d471b472980281b67dcb7b Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.717640 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.717639 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" event={"ID":"66de2d9c-8515-4ba8-ae01-100732a4e407","Type":"ContainerStarted","Data":"15a73937653ecd21f0b4930f89070e021ee7a2f306d60666388239941d3ccaec"} Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.718997 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d3ba584_a688_4fed_bba4_4a1b5e374703.slice/crio-72277c75fdec6901c14c180c6e032cef4810912872cbca73ae6ce4548a4e6765 WatchSource:0}: Error finding container 72277c75fdec6901c14c180c6e032cef4810912872cbca73ae6ce4548a4e6765: Status 404 returned error can't find the container with id 72277c75fdec6901c14c180c6e032cef4810912872cbca73ae6ce4548a4e6765 Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.721753 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf308111_7e73_4d70_b8f5_7c31b6ed3cda.slice/crio-2380c88e638d4d8f7eb681683a074a996f2227d5dacada13d9ab4badbc932bde WatchSource:0}: Error finding container 2380c88e638d4d8f7eb681683a074a996f2227d5dacada13d9ab4badbc932bde: Status 404 returned error can't find the container with id 2380c88e638d4d8f7eb681683a074a996f2227d5dacada13d9ab4badbc932bde Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.722265 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" event={"ID":"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8","Type":"ContainerStarted","Data":"311113ba131fdb8a857b5341387867b26b591f5170946899d94dc63e38a03fac"} Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.723772 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17dcd89e_7e17_408c_b7d8_806425dafbb9.slice/crio-f4dab0dfd804b7f17dfa32c53ccd8b54351503f5ff16bf2f5c8cd60b18170ff9 WatchSource:0}: Error finding container f4dab0dfd804b7f17dfa32c53ccd8b54351503f5ff16bf2f5c8cd60b18170ff9: Status 404 returned error can't find the container with id f4dab0dfd804b7f17dfa32c53ccd8b54351503f5ff16bf2f5c8cd60b18170ff9 Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.724139 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" event={"ID":"4e308683-ec1e-400b-8503-e671d8ac4589","Type":"ContainerStarted","Data":"854ea3b4d550805174ac397ecd4af381674fd11e7b5c05ea7e37680b0cd250f6"} Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.725558 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ab548fc_3aaf_4aef_9982_64bb1c243e90.slice/crio-610f814fd44ea74c5bc1c3f4d42cd9a5dc5432f13b05acb77481e4377c04d40f WatchSource:0}: Error finding container 610f814fd44ea74c5bc1c3f4d42cd9a5dc5432f13b05acb77481e4377c04d40f: Status 404 returned error can't find the container with id 610f814fd44ea74c5bc1c3f4d42cd9a5dc5432f13b05acb77481e4377c04d40f Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.725604 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.733482 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.734715 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" event={"ID":"dd618c39-c17a-40d9-add1-f87a6f646241","Type":"ContainerStarted","Data":"53266cde319546d78ff1e41e87d82bd34f57bed41a8c25dfdee6b6a15ede6da7"} Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.735552 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" event={"ID":"fd597d15-c58a-4569-9482-70a16d2a3133","Type":"ContainerStarted","Data":"b847e6029106cfae9c2409bdf36a094614ae0758e020958b81a1751b7ada567c"} Dec 06 06:55:29 crc kubenswrapper[4945]: W1206 06:55:29.736510 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b5df8d0_7bf5_478b_adb7_eb7291da8935.slice/crio-a35e442a12c312622f7c657f0a160de4f68bbd0d1843933fc649554705ce58a4 WatchSource:0}: Error finding container a35e442a12c312622f7c657f0a160de4f68bbd0d1843933fc649554705ce58a4: Status 404 returned error can't find the container with id a35e442a12c312622f7c657f0a160de4f68bbd0d1843933fc649554705ce58a4 Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.736864 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" event={"ID":"957f12af-ad21-4c4a-97a2-ab1b824aa5a0","Type":"ContainerStarted","Data":"62a2d51b6a9d7d7381fc1894e81e088920deb46b61702c870bc5daafd1d9d359"} Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.741602 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" event={"ID":"e06b2844-0600-4df6-bf5e-749ed15e596a","Type":"ContainerStarted","Data":"cb9c43d89b3347de5f75ba68e4a1cf36a3c36bcdfa7f71d728f266f34ff2b233"} Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.747674 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.756125 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.773790 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.775189 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.275160234 +0000 UTC m=+143.730021278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.787888 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.795914 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.803880 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.875827 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.876636 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.37658485 +0000 UTC m=+143.831445964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.896799 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64"] Dec 06 06:55:29 crc kubenswrapper[4945]: I1206 06:55:29.977215 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:29 crc kubenswrapper[4945]: E1206 06:55:29.977601 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.477566855 +0000 UTC m=+143.932428059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.085110 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.085715 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.585700276 +0000 UTC m=+144.040561320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.103365 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6"] Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.121538 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf"] Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.187447 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.187837 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.68781773 +0000 UTC m=+144.142678774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.289971 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.290409 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.790387255 +0000 UTC m=+144.245248299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.365333 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm"] Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.376597 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5"] Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.392013 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.392260 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.892222722 +0000 UTC m=+144.347083766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.392714 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.393231 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.893222138 +0000 UTC m=+144.348083182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.408447 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f"] Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.434767 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz"] Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.494364 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.494752 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.994712215 +0000 UTC m=+144.449573269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.496485 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.497167 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:30.997149939 +0000 UTC m=+144.452010993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.523335 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx"] Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.536113 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq"] Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.539988 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-j2tjw"] Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.559242 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d"] Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.597659 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.597860 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.097829116 +0000 UTC m=+144.552690160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.598480 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.598976 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.098951115 +0000 UTC m=+144.553812229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: W1206 06:55:30.670078 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd69d15e_5487_4c83_8f34_175811403e6a.slice/crio-017c33d2999c3f4705fb2d6dec0a1dfc9411b9806bcd267fb7dd9e608a4343e4 WatchSource:0}: Error finding container 017c33d2999c3f4705fb2d6dec0a1dfc9411b9806bcd267fb7dd9e608a4343e4: Status 404 returned error can't find the container with id 017c33d2999c3f4705fb2d6dec0a1dfc9411b9806bcd267fb7dd9e608a4343e4 Dec 06 06:55:30 crc kubenswrapper[4945]: W1206 06:55:30.670842 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode86a5672_d1fa_4706_a073_90193128f097.slice/crio-6255589ef5189cf2e9196e98986a1e95e3321fcc7e76cb8ea9ea57441bf9a61a WatchSource:0}: Error finding container 6255589ef5189cf2e9196e98986a1e95e3321fcc7e76cb8ea9ea57441bf9a61a: Status 404 returned error can't find the container with id 6255589ef5189cf2e9196e98986a1e95e3321fcc7e76cb8ea9ea57441bf9a61a Dec 06 06:55:30 crc kubenswrapper[4945]: W1206 06:55:30.672696 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fc7fa8b_727b_439a_bf88_7210f1387f6d.slice/crio-8ec97afda6aa480052291fe8129002aa1c84e9410a3801bd606d7340987e3f41 WatchSource:0}: Error finding container 8ec97afda6aa480052291fe8129002aa1c84e9410a3801bd606d7340987e3f41: Status 404 returned error can't find the container with id 8ec97afda6aa480052291fe8129002aa1c84e9410a3801bd606d7340987e3f41 Dec 06 06:55:30 crc kubenswrapper[4945]: W1206 06:55:30.674834 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23ec3b33_0492_49bd_b1df_aa5aaac01315.slice/crio-0c9e71ce105f0e54cdf7c3830491180b1d41796e106aa03b6768631db90caadd WatchSource:0}: Error finding container 0c9e71ce105f0e54cdf7c3830491180b1d41796e106aa03b6768631db90caadd: Status 404 returned error can't find the container with id 0c9e71ce105f0e54cdf7c3830491180b1d41796e106aa03b6768631db90caadd Dec 06 06:55:30 crc kubenswrapper[4945]: W1206 06:55:30.680500 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9ef2f41_8ec0_48fe_8b43_a2d230b60392.slice/crio-658936adedcaa8e8e41c6aa67e7a2706bc40690493690a306c004e9c840e0707 WatchSource:0}: Error finding container 658936adedcaa8e8e41c6aa67e7a2706bc40690493690a306c004e9c840e0707: Status 404 returned error can't find the container with id 658936adedcaa8e8e41c6aa67e7a2706bc40690493690a306c004e9c840e0707 Dec 06 06:55:30 crc kubenswrapper[4945]: W1206 06:55:30.684963 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6b25270_c23f_4cbb_b697_82ead103ae98.slice/crio-0d8ee4b5c645fc9e795a30c77700bd8f29750a0eb3ba4dc7f50a15535fc57049 WatchSource:0}: Error finding container 0d8ee4b5c645fc9e795a30c77700bd8f29750a0eb3ba4dc7f50a15535fc57049: Status 404 returned error can't find the container with id 0d8ee4b5c645fc9e795a30c77700bd8f29750a0eb3ba4dc7f50a15535fc57049 Dec 06 06:55:30 crc kubenswrapper[4945]: W1206 06:55:30.686526 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11106fdc_adfc_4dd8_8011_599912030ba5.slice/crio-82ab7ddbfa33fb5f3b3e2130b1619a662881349c8fa4ca986145bd4c7e45b5b3 WatchSource:0}: Error finding container 82ab7ddbfa33fb5f3b3e2130b1619a662881349c8fa4ca986145bd4c7e45b5b3: Status 404 returned error can't find the container with id 82ab7ddbfa33fb5f3b3e2130b1619a662881349c8fa4ca986145bd4c7e45b5b3 Dec 06 06:55:30 crc kubenswrapper[4945]: W1206 06:55:30.689096 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50e16132_af33_4dbf_b4f4_a9b0c8554e24.slice/crio-72e210eea0aa09037ed3274ef7f7dc35f84fc3eb838760a9f0e30f787c77cb04 WatchSource:0}: Error finding container 72e210eea0aa09037ed3274ef7f7dc35f84fc3eb838760a9f0e30f787c77cb04: Status 404 returned error can't find the container with id 72e210eea0aa09037ed3274ef7f7dc35f84fc3eb838760a9f0e30f787c77cb04 Dec 06 06:55:30 crc kubenswrapper[4945]: W1206 06:55:30.695015 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7adf125b_8ab8_42b0_9d4d_c38aacbc1a17.slice/crio-08b6516eb6f53a1d90b5cee3c523fa8ef728afcff6183de5dde9dd4d8cc43e80 WatchSource:0}: Error finding container 08b6516eb6f53a1d90b5cee3c523fa8ef728afcff6183de5dde9dd4d8cc43e80: Status 404 returned error can't find the container with id 08b6516eb6f53a1d90b5cee3c523fa8ef728afcff6183de5dde9dd4d8cc43e80 Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.699243 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.699545 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.199518088 +0000 UTC m=+144.654379132 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.699706 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.700713 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.20070481 +0000 UTC m=+144.655565844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.749184 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" event={"ID":"5d3ba584-a688-4fed-bba4-4a1b5e374703","Type":"ContainerStarted","Data":"72277c75fdec6901c14c180c6e032cef4810912872cbca73ae6ce4548a4e6765"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.751966 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" event={"ID":"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17","Type":"ContainerStarted","Data":"08b6516eb6f53a1d90b5cee3c523fa8ef728afcff6183de5dde9dd4d8cc43e80"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.753392 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" event={"ID":"df308111-7e73-4d70-b8f5-7c31b6ed3cda","Type":"ContainerStarted","Data":"2380c88e638d4d8f7eb681683a074a996f2227d5dacada13d9ab4badbc932bde"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.774637 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" event={"ID":"3254c754-c3b0-41c7-93d7-f52776056d30","Type":"ContainerStarted","Data":"35fb73f0e10834299201127e329ad805b6d4e6654c6ac474c30a8acf91c54638"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.780058 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" event={"ID":"11106fdc-adfc-4dd8-8011-599912030ba5","Type":"ContainerStarted","Data":"82ab7ddbfa33fb5f3b3e2130b1619a662881349c8fa4ca986145bd4c7e45b5b3"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.783499 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" event={"ID":"5fc7fa8b-727b-439a-bf88-7210f1387f6d","Type":"ContainerStarted","Data":"8ec97afda6aa480052291fe8129002aa1c84e9410a3801bd606d7340987e3f41"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.806199 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.806330 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.306306625 +0000 UTC m=+144.761167659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.806662 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.807177 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.307153437 +0000 UTC m=+144.762014481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.816542 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" event={"ID":"2d9fa11b-0c16-4915-b12c-8cc917d82fb7","Type":"ContainerStarted","Data":"8aa370e68b73131f7b3e603cc794ba8150411acf1190a6eb688d07b552b1fce7"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.824877 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" event={"ID":"7d28b3c8-4c1d-49b4-bc26-47cdb9867e53","Type":"ContainerStarted","Data":"cd7e583cc4cebbf82abbf9b36862237ea0ef970397d471b472980281b67dcb7b"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.846595 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" event={"ID":"dd69d15e-5487-4c83-8f34-175811403e6a","Type":"ContainerStarted","Data":"017c33d2999c3f4705fb2d6dec0a1dfc9411b9806bcd267fb7dd9e608a4343e4"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.896143 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" event={"ID":"17dcd89e-7e17-408c-b7d8-806425dafbb9","Type":"ContainerStarted","Data":"f4dab0dfd804b7f17dfa32c53ccd8b54351503f5ff16bf2f5c8cd60b18170ff9"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.907424 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:30 crc kubenswrapper[4945]: E1206 06:55:30.907826 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.407807063 +0000 UTC m=+144.862668097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.966625 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" event={"ID":"c0b86555-314f-41f4-8a5d-5aeb356fc01f","Type":"ContainerStarted","Data":"2fa463cf3987368a3d81db05ace07a4fe151a30a2cb63bd428f97cef5caa9010"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.966953 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" event={"ID":"5ab548fc-3aaf-4aef-9982-64bb1c243e90","Type":"ContainerStarted","Data":"610f814fd44ea74c5bc1c3f4d42cd9a5dc5432f13b05acb77481e4377c04d40f"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.966967 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" event={"ID":"23ec3b33-0492-49bd-b1df-aa5aaac01315","Type":"ContainerStarted","Data":"0c9e71ce105f0e54cdf7c3830491180b1d41796e106aa03b6768631db90caadd"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.966979 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" event={"ID":"b6b25270-c23f-4cbb-b697-82ead103ae98","Type":"ContainerStarted","Data":"0d8ee4b5c645fc9e795a30c77700bd8f29750a0eb3ba4dc7f50a15535fc57049"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.967237 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" event={"ID":"e86a5672-d1fa-4706-a073-90193128f097","Type":"ContainerStarted","Data":"6255589ef5189cf2e9196e98986a1e95e3321fcc7e76cb8ea9ea57441bf9a61a"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.968255 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-j2tjw" event={"ID":"50e16132-af33-4dbf-b4f4-a9b0c8554e24","Type":"ContainerStarted","Data":"72e210eea0aa09037ed3274ef7f7dc35f84fc3eb838760a9f0e30f787c77cb04"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.971933 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lpl6d" event={"ID":"5bfd04a8-aad7-482a-9e68-f74c9cff00ba","Type":"ContainerStarted","Data":"3e56926cce01c3e3623f706c7db3ebe2ab5031713d8de97edae012f7b3b873d6"} Dec 06 06:55:30 crc kubenswrapper[4945]: I1206 06:55:30.982225 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" event={"ID":"c9ef2f41-8ec0-48fe-8b43-a2d230b60392","Type":"ContainerStarted","Data":"658936adedcaa8e8e41c6aa67e7a2706bc40690493690a306c004e9c840e0707"} Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.002732 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bl6dw" event={"ID":"28449122-a88f-42b7-8d19-c1875c7b329d","Type":"ContainerStarted","Data":"677d267745fbb0de7cb79dc2490d9abd0277b6f2eada36afe03bfa7cb8da6f15"} Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.009832 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.010434 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.510414199 +0000 UTC m=+144.965275253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.015118 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv" event={"ID":"5b5df8d0-7bf5-478b-adb7-eb7291da8935","Type":"ContainerStarted","Data":"a35e442a12c312622f7c657f0a160de4f68bbd0d1843933fc649554705ce58a4"} Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.111524 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.112015 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.611987269 +0000 UTC m=+145.066848323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.214149 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.214711 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.714690889 +0000 UTC m=+145.169551933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.315699 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.315925 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.815889609 +0000 UTC m=+145.270750653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.316447 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.316846 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.816837553 +0000 UTC m=+145.271698597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.417386 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.417968 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:31.917947301 +0000 UTC m=+145.372808345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.511595 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kngc2"] Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.515020 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6"] Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.522188 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mgmz8"] Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.522975 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.523361 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.023346601 +0000 UTC m=+145.478207645 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.598546 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cf8e342_cc9f_4017_87de_ea4f1dcaa396.slice/crio-2fb08a59c85c7bbf89a766d0f0ed76a1f0103050012596abfeb509817e3a0ff2 WatchSource:0}: Error finding container 2fb08a59c85c7bbf89a766d0f0ed76a1f0103050012596abfeb509817e3a0ff2: Status 404 returned error can't find the container with id 2fb08a59c85c7bbf89a766d0f0ed76a1f0103050012596abfeb509817e3a0ff2 Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.624046 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.625603 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.12448465 +0000 UTC m=+145.579345694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.642119 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rnq6j"] Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.644476 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf"] Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.657866 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77e91d35_8908_44ee_b5b8_86cdb9b46726.slice/crio-bba20c6e660345934d56d582e2130203888efe55768ce3b51342973461202d1a WatchSource:0}: Error finding container bba20c6e660345934d56d582e2130203888efe55768ce3b51342973461202d1a: Status 404 returned error can't find the container with id bba20c6e660345934d56d582e2130203888efe55768ce3b51342973461202d1a Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.709496 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e0a6aac_8db4_45b0_9d8d_bbd99962bd3c.slice/crio-e5af57fd4018ed6052615cd4aff5d91ed6ac54dcda99289a903daeea9e118bbb WatchSource:0}: Error finding container e5af57fd4018ed6052615cd4aff5d91ed6ac54dcda99289a903daeea9e118bbb: Status 404 returned error can't find the container with id e5af57fd4018ed6052615cd4aff5d91ed6ac54dcda99289a903daeea9e118bbb Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.711971 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a24d873_f77d_43d9_a402_b7a3d9194496.slice/crio-48dc03749c61be8e42d4b7c2baecd2ec9053d407d1e6ca1b38e3d12e861ef6db WatchSource:0}: Error finding container 48dc03749c61be8e42d4b7c2baecd2ec9053d407d1e6ca1b38e3d12e861ef6db: Status 404 returned error can't find the container with id 48dc03749c61be8e42d4b7c2baecd2ec9053d407d1e6ca1b38e3d12e861ef6db Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.727507 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f"] Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.730017 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5b5kg"] Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.735477 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz"] Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.737241 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8skx"] Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.739129 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.739524 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dvn5j"] Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.741806 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6"] Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.742501 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.239800079 +0000 UTC m=+145.694661123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.755411 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8194654b_ee85_4388_bc86_633437991bab.slice/crio-3e5cb82bbdf7fb69ab681259d841c6884504691802646576db926c3a3ed0d5a0 WatchSource:0}: Error finding container 3e5cb82bbdf7fb69ab681259d841c6884504691802646576db926c3a3ed0d5a0: Status 404 returned error can't find the container with id 3e5cb82bbdf7fb69ab681259d841c6884504691802646576db926c3a3ed0d5a0 Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.761554 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cf4n5"] Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.769185 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d7466cb_add7_436c_8641_0bf1bda1afbd.slice/crio-a270ceb9f171a5167da5f367a842e4ae81ac5fb2ff9ab940f7cb7bbd60dbce8f WatchSource:0}: Error finding container a270ceb9f171a5167da5f367a842e4ae81ac5fb2ff9ab940f7cb7bbd60dbce8f: Status 404 returned error can't find the container with id a270ceb9f171a5167da5f367a842e4ae81ac5fb2ff9ab940f7cb7bbd60dbce8f Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.769584 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb79525cf_ee6a_4a3b_a83d_4df325832a72.slice/crio-5302dee1d4dc5beb4c03d7ae64b5330b6f320a251f2340de63efbe4a21fa76f6 WatchSource:0}: Error finding container 5302dee1d4dc5beb4c03d7ae64b5330b6f320a251f2340de63efbe4a21fa76f6: Status 404 returned error can't find the container with id 5302dee1d4dc5beb4c03d7ae64b5330b6f320a251f2340de63efbe4a21fa76f6 Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.770146 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7045282_9b6c_4821_aa59_d0f45f65682b.slice/crio-ed96923f9b4f301ac0d4cbc3167cd401125e9eea47ee0cfae744389419ee59b8 WatchSource:0}: Error finding container ed96923f9b4f301ac0d4cbc3167cd401125e9eea47ee0cfae744389419ee59b8: Status 404 returned error can't find the container with id ed96923f9b4f301ac0d4cbc3167cd401125e9eea47ee0cfae744389419ee59b8 Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.774908 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb5dc394_7097_4efd_893a_5869e136d910.slice/crio-f3ed48dea56f470d77e758a5399a1014f33da3b625ea81fc13519bb4dd32d8c8 WatchSource:0}: Error finding container f3ed48dea56f470d77e758a5399a1014f33da3b625ea81fc13519bb4dd32d8c8: Status 404 returned error can't find the container with id f3ed48dea56f470d77e758a5399a1014f33da3b625ea81fc13519bb4dd32d8c8 Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.777739 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9ad8684_da2b_4ac8_ab89_a6eb3ab47311.slice/crio-c75b2c5b8a50ccc1da5d52842dc3649b27a3fadac6288bd1596fb979beca49e9 WatchSource:0}: Error finding container c75b2c5b8a50ccc1da5d52842dc3649b27a3fadac6288bd1596fb979beca49e9: Status 404 returned error can't find the container with id c75b2c5b8a50ccc1da5d52842dc3649b27a3fadac6288bd1596fb979beca49e9 Dec 06 06:55:31 crc kubenswrapper[4945]: W1206 06:55:31.795401 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9bf5fde_fb1b_4e2f_ab6f_97aef4fa9c88.slice/crio-dcccb9f4ec7e31f564cf05d4bbf42c98204505e38d54c33bc02b7f8576b4160b WatchSource:0}: Error finding container dcccb9f4ec7e31f564cf05d4bbf42c98204505e38d54c33bc02b7f8576b4160b: Status 404 returned error can't find the container with id dcccb9f4ec7e31f564cf05d4bbf42c98204505e38d54c33bc02b7f8576b4160b Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.840954 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.841384 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.341359879 +0000 UTC m=+145.796220923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:31 crc kubenswrapper[4945]: I1206 06:55:31.942517 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:31 crc kubenswrapper[4945]: E1206 06:55:31.942926 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.442906658 +0000 UTC m=+145.897767702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.044611 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.045344 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.545275178 +0000 UTC m=+146.000136222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.146647 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.147211 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.647196887 +0000 UTC m=+146.102057931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.147590 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" event={"ID":"17dcd89e-7e17-408c-b7d8-806425dafbb9","Type":"ContainerStarted","Data":"2c58ccc65e1fa4b56c144106aee55969f5c407735da5a2a7297cda1d4f3097c0"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.185473 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" event={"ID":"4e308683-ec1e-400b-8503-e671d8ac4589","Type":"ContainerStarted","Data":"8e726e76165215a17fe8c50e650c303c138066e5f03649bdaf57f2837f74a6c0"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.193744 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" event={"ID":"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311","Type":"ContainerStarted","Data":"c75b2c5b8a50ccc1da5d52842dc3649b27a3fadac6288bd1596fb979beca49e9"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.231679 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sg6tn" podStartSLOduration=124.231648718 podStartE2EDuration="2m4.231648718s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.175065547 +0000 UTC m=+145.629926591" watchObservedRunningTime="2025-12-06 06:55:32.231648718 +0000 UTC m=+145.686509762" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.244944 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.248470 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.248492 4945 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-kbjdx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.248557 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" podUID="b6b25270-c23f-4cbb-b697-82ead103ae98" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.248889 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.748738116 +0000 UTC m=+146.203599160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.248941 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.249471 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.749445814 +0000 UTC m=+146.204307038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.250495 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" event={"ID":"a7045282-9b6c-4821-aa59-d0f45f65682b","Type":"ContainerStarted","Data":"ed96923f9b4f301ac0d4cbc3167cd401125e9eea47ee0cfae744389419ee59b8"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.252429 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" event={"ID":"9a24d873-f77d-43d9-a402-b7a3d9194496","Type":"ContainerStarted","Data":"48dc03749c61be8e42d4b7c2baecd2ec9053d407d1e6ca1b38e3d12e861ef6db"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.256686 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" event={"ID":"8d7466cb-add7-436c-8641-0bf1bda1afbd","Type":"ContainerStarted","Data":"a270ceb9f171a5167da5f367a842e4ae81ac5fb2ff9ab940f7cb7bbd60dbce8f"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.261488 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" event={"ID":"e06b2844-0600-4df6-bf5e-749ed15e596a","Type":"ContainerStarted","Data":"1462cbd3aeabfb5e5a0b333916b901cc13097d082dfc56cb4c534f6b40fc7d5b"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.262494 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.278946 4945 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-x95jj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.279021 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" podUID="e06b2844-0600-4df6-bf5e-749ed15e596a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.280585 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" podStartSLOduration=123.28056904 podStartE2EDuration="2m3.28056904s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.280348874 +0000 UTC m=+145.735209938" watchObservedRunningTime="2025-12-06 06:55:32.28056904 +0000 UTC m=+145.735430084" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.281036 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" podStartSLOduration=124.281032652 podStartE2EDuration="2m4.281032652s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.233638451 +0000 UTC m=+145.688499505" watchObservedRunningTime="2025-12-06 06:55:32.281032652 +0000 UTC m=+145.735893696" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.308435 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kngc2" event={"ID":"77e91d35-8908-44ee-b5b8-86cdb9b46726","Type":"ContainerStarted","Data":"bba20c6e660345934d56d582e2130203888efe55768ce3b51342973461202d1a"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.328863 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" event={"ID":"cb5dc394-7097-4efd-893a-5869e136d910","Type":"ContainerStarted","Data":"f3ed48dea56f470d77e758a5399a1014f33da3b625ea81fc13519bb4dd32d8c8"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.330634 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" event={"ID":"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88","Type":"ContainerStarted","Data":"dcccb9f4ec7e31f564cf05d4bbf42c98204505e38d54c33bc02b7f8576b4160b"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.336468 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" event={"ID":"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8","Type":"ContainerStarted","Data":"8bf6bdbd8a892359c6989f77af3ac4e2a5192aa26f7e00cf316832915a393fa9"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.342479 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" event={"ID":"8194654b-ee85-4388-bc86-633437991bab","Type":"ContainerStarted","Data":"3e5cb82bbdf7fb69ab681259d841c6884504691802646576db926c3a3ed0d5a0"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.344669 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" event={"ID":"9cf8e342-cc9f-4017-87de-ea4f1dcaa396","Type":"ContainerStarted","Data":"2fb08a59c85c7bbf89a766d0f0ed76a1f0103050012596abfeb509817e3a0ff2"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.350853 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.352321 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.852267977 +0000 UTC m=+146.307129021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.377798 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" event={"ID":"dd618c39-c17a-40d9-add1-f87a6f646241","Type":"ContainerStarted","Data":"6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.378752 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.380015 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" podStartSLOduration=124.379998483 podStartE2EDuration="2m4.379998483s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.318080172 +0000 UTC m=+145.772941216" watchObservedRunningTime="2025-12-06 06:55:32.379998483 +0000 UTC m=+145.834859527" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.381114 4945 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-xrvfp container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.381170 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" podUID="dd618c39-c17a-40d9-add1-f87a6f646241" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.422892 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" event={"ID":"91c9713c-0fd9-41fb-a4dd-b4d5bc691131","Type":"ContainerStarted","Data":"b600c0546c61e96137e1a4196cae5dc631c00def8b80c15f7a8c58420438fcd7"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.425378 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" event={"ID":"66de2d9c-8515-4ba8-ae01-100732a4e407","Type":"ContainerStarted","Data":"d8e15426a3a41d570db287a89f1da6e6d303a2774fdb21661af4be6dbb6996e0"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.427199 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" event={"ID":"b79525cf-ee6a-4a3b-a83d-4df325832a72","Type":"ContainerStarted","Data":"5302dee1d4dc5beb4c03d7ae64b5330b6f320a251f2340de63efbe4a21fa76f6"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.440357 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" podStartSLOduration=123.440334503 podStartE2EDuration="2m3.440334503s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.433931655 +0000 UTC m=+145.888792699" watchObservedRunningTime="2025-12-06 06:55:32.440334503 +0000 UTC m=+145.895195547" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.444301 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mgmz8" event={"ID":"8708c768-ebe0-4070-b7b6-249649beb4f7","Type":"ContainerStarted","Data":"c9795672b76404ae7340b610fa7c646f780f5d6ff78e1f805ad6ff67c9c3e4f5"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.444707 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.458225 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.459519 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:32.959504875 +0000 UTC m=+146.414365919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.468870 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" podStartSLOduration=124.46885045 podStartE2EDuration="2m4.46885045s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.466912869 +0000 UTC m=+145.921773923" watchObservedRunningTime="2025-12-06 06:55:32.46885045 +0000 UTC m=+145.923711504" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.469871 4945 patch_prober.go:28] interesting pod/console-operator-58897d9998-mgmz8 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/readyz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.469919 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-mgmz8" podUID="8708c768-ebe0-4070-b7b6-249649beb4f7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/readyz\": dial tcp 10.217.0.31:8443: connect: connection refused" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.509705 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4c2s9" podStartSLOduration=124.509677089 podStartE2EDuration="2m4.509677089s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.494802689 +0000 UTC m=+145.949663733" watchObservedRunningTime="2025-12-06 06:55:32.509677089 +0000 UTC m=+145.964538133" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.522908 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bl6dw" podStartSLOduration=6.522887265 podStartE2EDuration="6.522887265s" podCreationTimestamp="2025-12-06 06:55:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.521721024 +0000 UTC m=+145.976582068" watchObservedRunningTime="2025-12-06 06:55:32.522887265 +0000 UTC m=+145.977748309" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.523964 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rnq6j" event={"ID":"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c","Type":"ContainerStarted","Data":"e5af57fd4018ed6052615cd4aff5d91ed6ac54dcda99289a903daeea9e118bbb"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.545217 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" event={"ID":"5ab548fc-3aaf-4aef-9982-64bb1c243e90","Type":"ContainerStarted","Data":"2efdc6dbd597c2dde1dcbc5308765c11a3280d05c0e3e1b21bdc5a4f45ee9d7f"} Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.558150 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-mgmz8" podStartSLOduration=124.55812327699999 podStartE2EDuration="2m4.558123277s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.557478251 +0000 UTC m=+146.012339305" watchObservedRunningTime="2025-12-06 06:55:32.558123277 +0000 UTC m=+146.012984321" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.560992 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.565292 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:33.065244524 +0000 UTC m=+146.520105568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.609820 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-j2tjw" podStartSLOduration=124.60978441 podStartE2EDuration="2m4.60978441s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.585009652 +0000 UTC m=+146.039870696" watchObservedRunningTime="2025-12-06 06:55:32.60978441 +0000 UTC m=+146.064645454" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.628114 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-bs8n9" podStartSLOduration=124.628079099 podStartE2EDuration="2m4.628079099s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.618129069 +0000 UTC m=+146.072990113" watchObservedRunningTime="2025-12-06 06:55:32.628079099 +0000 UTC m=+146.082940143" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.661084 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.663141 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.665457 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:33.165440168 +0000 UTC m=+146.620301212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.669625 4945 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-vj29d container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.669666 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" podUID="7adf125b-8ab8-42b0-9d4d-c38aacbc1a17" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.696934 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" podStartSLOduration=124.696912482 podStartE2EDuration="2m4.696912482s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.696033899 +0000 UTC m=+146.150894943" watchObservedRunningTime="2025-12-06 06:55:32.696912482 +0000 UTC m=+146.151773526" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.734634 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" podStartSLOduration=123.734606279 podStartE2EDuration="2m3.734606279s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:32.726336542 +0000 UTC m=+146.181197596" watchObservedRunningTime="2025-12-06 06:55:32.734606279 +0000 UTC m=+146.189467323" Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.766571 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.766803 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:33.266782701 +0000 UTC m=+146.721643745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.767296 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.769273 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:33.269260986 +0000 UTC m=+146.724122030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.868867 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.869673 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:33.369648145 +0000 UTC m=+146.824509199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:32 crc kubenswrapper[4945]: I1206 06:55:32.971485 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:32 crc kubenswrapper[4945]: E1206 06:55:32.972233 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:33.472219341 +0000 UTC m=+146.927080375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.078978 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:33 crc kubenswrapper[4945]: E1206 06:55:33.079540 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:33.579516851 +0000 UTC m=+147.034377895 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.181085 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:33 crc kubenswrapper[4945]: E1206 06:55:33.182203 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:33.682183979 +0000 UTC m=+147.137045033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.295700 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:33 crc kubenswrapper[4945]: E1206 06:55:33.296250 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:33.796227015 +0000 UTC m=+147.251088059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.405045 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:33 crc kubenswrapper[4945]: E1206 06:55:33.406079 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:33.906063242 +0000 UTC m=+147.360924286 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.515601 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:33 crc kubenswrapper[4945]: E1206 06:55:33.515998 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.01597915 +0000 UTC m=+147.470840194 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.618304 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:33 crc kubenswrapper[4945]: E1206 06:55:33.618685 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.118669489 +0000 UTC m=+147.573530533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.686785 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" event={"ID":"7d28b3c8-4c1d-49b4-bc26-47cdb9867e53","Type":"ContainerStarted","Data":"3b79ccd367291d62bae05e52d29ab10055e6930cc26c1682dcb4b9319ea5e5fb"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.687320 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" event={"ID":"7d28b3c8-4c1d-49b4-bc26-47cdb9867e53","Type":"ContainerStarted","Data":"5166b7e2edcd7c928c46ef69060d0d8681419ae2ca0989cea4090bc709d3d7ff"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.713670 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" event={"ID":"b79525cf-ee6a-4a3b-a83d-4df325832a72","Type":"ContainerStarted","Data":"b86eeead52fc3aada55cacf66cc2aa37e049b72abc36f786ef9fe0baf94e2db8"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.719836 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:33 crc kubenswrapper[4945]: E1206 06:55:33.720271 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.220251239 +0000 UTC m=+147.675112283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.747014 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" event={"ID":"a9ad8684-da2b-4ac8-ab89-a6eb3ab47311","Type":"ContainerStarted","Data":"d5d7441397aa16d3cee7461f18c8ba13ca84b49061aaad56cc9842e6f585aa8d"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.779636 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" event={"ID":"5d3ba584-a688-4fed-bba4-4a1b5e374703","Type":"ContainerStarted","Data":"2cd46bbd222ca1dbfe852cec1b675e1a1f50980d0fd412c8bd3db268b229bb74"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.779704 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" event={"ID":"5d3ba584-a688-4fed-bba4-4a1b5e374703","Type":"ContainerStarted","Data":"562a74f6e116292bb8c12eee97583cf7de4427a651fe831cc0d59e9af93c0123"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.780804 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-pfkqs" podStartSLOduration=124.780793733 podStartE2EDuration="2m4.780793733s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:33.77915954 +0000 UTC m=+147.234020594" watchObservedRunningTime="2025-12-06 06:55:33.780793733 +0000 UTC m=+147.235654777" Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.801727 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" event={"ID":"b6b25270-c23f-4cbb-b697-82ead103ae98","Type":"ContainerStarted","Data":"0cbb8f42524cb47d26b4a8cb3699636e33775e624c4a695a072cdfe48ad3934f"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.818572 4945 generic.go:334] "Generic (PLEG): container finished" podID="fd597d15-c58a-4569-9482-70a16d2a3133" containerID="1a801b08804eb0116f3bff9979f9444a4a363de22e0da0013f5c679005faa394" exitCode=0 Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.818704 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" event={"ID":"fd597d15-c58a-4569-9482-70a16d2a3133","Type":"ContainerDied","Data":"1a801b08804eb0116f3bff9979f9444a4a363de22e0da0013f5c679005faa394"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.825043 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:33 crc kubenswrapper[4945]: E1206 06:55:33.827066 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.327043694 +0000 UTC m=+147.781904928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.864629 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rnq6j" event={"ID":"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c","Type":"ContainerStarted","Data":"ec4798d12b585c52942f80447b85cf4fc4a653e475043eb101125f02f9866a58"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.865575 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.880184 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kbjdx" Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.884514 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kngc2" event={"ID":"77e91d35-8908-44ee-b5b8-86cdb9b46726","Type":"ContainerStarted","Data":"08d68294929617e9bbe034a1dd9a99f619f6b105914a42ccb936c20302e06c92"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.897590 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4wqcz" podStartSLOduration=124.897563951 podStartE2EDuration="2m4.897563951s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:33.852751447 +0000 UTC m=+147.307612491" watchObservedRunningTime="2025-12-06 06:55:33.897563951 +0000 UTC m=+147.352425005" Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.897952 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5g55f" podStartSLOduration=124.897948171 podStartE2EDuration="2m4.897948171s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:33.891665647 +0000 UTC m=+147.346526681" watchObservedRunningTime="2025-12-06 06:55:33.897948171 +0000 UTC m=+147.352809215" Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.930457 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:33 crc kubenswrapper[4945]: E1206 06:55:33.932035 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.432009423 +0000 UTC m=+147.886870467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.933613 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" event={"ID":"9cf8e342-cc9f-4017-87de-ea4f1dcaa396","Type":"ContainerStarted","Data":"f921db093cb8b590a87b5b270b1522ece7fcb5ab97f2248ea8dad2b83442bb30"} Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.934795 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:33 crc kubenswrapper[4945]: I1206 06:55:33.974269 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" event={"ID":"cb5dc394-7097-4efd-893a-5869e136d910","Type":"ContainerStarted","Data":"7ed5b34fb1df9498926f6b835eddd7bd960a78c9442c81caaeb12e73d1eb9b0b"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.008465 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-j2tjw" event={"ID":"50e16132-af33-4dbf-b4f4-a9b0c8554e24","Type":"ContainerStarted","Data":"2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.010956 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" event={"ID":"8194654b-ee85-4388-bc86-633437991bab","Type":"ContainerStarted","Data":"9ba7e5446b907267bafbd3331dce2f84240d30f76052bf7b807b11347399251b"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.019429 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bl6dw" event={"ID":"28449122-a88f-42b7-8d19-c1875c7b329d","Type":"ContainerStarted","Data":"b0c828f7eea3eedaeada7dda98ddbd0a3b2bb47999dd13cc86dd1157bad54dc3"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.035169 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:34 crc kubenswrapper[4945]: E1206 06:55:34.036343 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.536324505 +0000 UTC m=+147.991185549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.037609 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rnq6j" podStartSLOduration=8.037589678 podStartE2EDuration="8.037589678s" podCreationTimestamp="2025-12-06 06:55:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:33.987559978 +0000 UTC m=+147.442421022" watchObservedRunningTime="2025-12-06 06:55:34.037589678 +0000 UTC m=+147.492450722" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.051389 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" event={"ID":"8d7466cb-add7-436c-8641-0bf1bda1afbd","Type":"ContainerStarted","Data":"aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.052600 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.060060 4945 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r8skx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.060139 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" podUID="8d7466cb-add7-436c-8641-0bf1bda1afbd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.094912 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" event={"ID":"11106fdc-adfc-4dd8-8011-599912030ba5","Type":"ContainerStarted","Data":"530a46883f12ad6ef08d1192e857c3272cee7404dd7197889089697d02e0a3d0"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.123265 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-kngc2" podStartSLOduration=8.123247311 podStartE2EDuration="8.123247311s" podCreationTimestamp="2025-12-06 06:55:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.072604085 +0000 UTC m=+147.527465129" watchObservedRunningTime="2025-12-06 06:55:34.123247311 +0000 UTC m=+147.578108355" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.123744 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-djxqd" podStartSLOduration=126.123738824 podStartE2EDuration="2m6.123738824s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.120782756 +0000 UTC m=+147.575643800" watchObservedRunningTime="2025-12-06 06:55:34.123738824 +0000 UTC m=+147.578599868" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.131684 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfp2d" event={"ID":"df308111-7e73-4d70-b8f5-7c31b6ed3cda","Type":"ContainerStarted","Data":"8f9d256aac9b2627e44de7314848d915fd6c6b09051af5fd2333d5b3cb1eb992"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.137380 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:34 crc kubenswrapper[4945]: E1206 06:55:34.139435 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.639403874 +0000 UTC m=+148.094265088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.177143 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lpl6d" event={"ID":"5bfd04a8-aad7-482a-9e68-f74c9cff00ba","Type":"ContainerStarted","Data":"5bd65ab6ca7b1153a6c2c2b7983f46c0941497df90018a335cfa4f88be445620"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.180050 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-lpl6d" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.183490 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-lpl6d container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.183606 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lpl6d" podUID="5bfd04a8-aad7-482a-9e68-f74c9cff00ba" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.192180 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2z8p6" podStartSLOduration=126.192146765 podStartE2EDuration="2m6.192146765s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.168040004 +0000 UTC m=+147.622901048" watchObservedRunningTime="2025-12-06 06:55:34.192146765 +0000 UTC m=+147.647007809" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.244712 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.270031 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" event={"ID":"a7045282-9b6c-4821-aa59-d0f45f65682b","Type":"ContainerStarted","Data":"68dacc1bc6afd071bb5481d07bac7d6d98dc1800653262f401f1a3a409f37de6"} Dec 06 06:55:34 crc kubenswrapper[4945]: E1206 06:55:34.286763 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.786744332 +0000 UTC m=+148.241605366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.313124 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" podStartSLOduration=125.313102772 podStartE2EDuration="2m5.313102772s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.269613633 +0000 UTC m=+147.724474677" watchObservedRunningTime="2025-12-06 06:55:34.313102772 +0000 UTC m=+147.767963816" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.314697 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" event={"ID":"91c9713c-0fd9-41fb-a4dd-b4d5bc691131","Type":"ContainerStarted","Data":"db19aab19bd880ac09584c36416dd0de688278453828a65b9a7641c0fefb2f64"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.314763 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" event={"ID":"91c9713c-0fd9-41fb-a4dd-b4d5bc691131","Type":"ContainerStarted","Data":"68031b60f16f4cc6dde45c36ac79627303e25f51653aafc6bf0b727696b6e857"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.328404 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mgmz8" event={"ID":"8708c768-ebe0-4070-b7b6-249649beb4f7","Type":"ContainerStarted","Data":"6f97e159a612f2fdd5a3ee5793114f5af9568dd8fe541228ab48819c0f0c741c"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.352720 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" podStartSLOduration=125.352696289 podStartE2EDuration="2m5.352696289s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.315732181 +0000 UTC m=+147.770593235" watchObservedRunningTime="2025-12-06 06:55:34.352696289 +0000 UTC m=+147.807557333" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.353429 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-clfbq" podStartSLOduration=126.353424368 podStartE2EDuration="2m6.353424368s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.3515702 +0000 UTC m=+147.806431254" watchObservedRunningTime="2025-12-06 06:55:34.353424368 +0000 UTC m=+147.808285412" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.371871 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:34 crc kubenswrapper[4945]: E1206 06:55:34.373311 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.873290568 +0000 UTC m=+148.328151622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.390343 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-5b5kg" podStartSLOduration=125.390323194 podStartE2EDuration="2m5.390323194s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.38938398 +0000 UTC m=+147.844245024" watchObservedRunningTime="2025-12-06 06:55:34.390323194 +0000 UTC m=+147.845184238" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.390964 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" event={"ID":"7adf125b-8ab8-42b0-9d4d-c38aacbc1a17","Type":"ContainerStarted","Data":"cfc091af9ba142d46ce95e7d596053bf55af161d0c2608e7cde9e5dccb02a006"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.432618 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vj29d" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.443651 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" event={"ID":"4e308683-ec1e-400b-8503-e671d8ac4589","Type":"ContainerStarted","Data":"981ca8ae96a5b4fed996412e70878cd9e6a77d01e751cbbd17d8580113c19317"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.446633 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-dvn5j" podStartSLOduration=125.446619118 podStartE2EDuration="2m5.446619118s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.445662723 +0000 UTC m=+147.900523767" watchObservedRunningTime="2025-12-06 06:55:34.446619118 +0000 UTC m=+147.901480182" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.458418 4945 generic.go:334] "Generic (PLEG): container finished" podID="3254c754-c3b0-41c7-93d7-f52776056d30" containerID="fa6736e4eb7fb28bd3f574148f2500b0e548adb7bb91c5c4a11a2ae454b90348" exitCode=0 Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.458552 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" event={"ID":"3254c754-c3b0-41c7-93d7-f52776056d30","Type":"ContainerDied","Data":"fa6736e4eb7fb28bd3f574148f2500b0e548adb7bb91c5c4a11a2ae454b90348"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.458584 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" event={"ID":"3254c754-c3b0-41c7-93d7-f52776056d30","Type":"ContainerStarted","Data":"75ca32e6e24fe4026fa64587dcc63aea1b67768ae9434641a6ed27dfc7d0c360"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.476321 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:34 crc kubenswrapper[4945]: E1206 06:55:34.477269 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:34.97724941 +0000 UTC m=+148.432110454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.507856 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" event={"ID":"dd69d15e-5487-4c83-8f34-175811403e6a","Type":"ContainerStarted","Data":"8e84fcd124ebd2f578467e7d60a053f2ae372ffa6fe728b7265fcaeb4dc10e96"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.518050 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dhk64" podStartSLOduration=126.518028808 podStartE2EDuration="2m6.518028808s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.517400302 +0000 UTC m=+147.972261536" watchObservedRunningTime="2025-12-06 06:55:34.518028808 +0000 UTC m=+147.972889852" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.546035 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" event={"ID":"2d9fa11b-0c16-4915-b12c-8cc917d82fb7","Type":"ContainerStarted","Data":"a0c2ee6a827f7abb5afd638a1240a9cb1b5178873dfb0fbf2e990dd5062f5fe8"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.546091 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" event={"ID":"2d9fa11b-0c16-4915-b12c-8cc917d82fb7","Type":"ContainerStarted","Data":"064406166bbe87f4385f7abb7fec9b4408021db61eb9274abe7401cd0cec5b24"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.565146 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-lpl6d" podStartSLOduration=126.565121612 podStartE2EDuration="2m6.565121612s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.555713625 +0000 UTC m=+148.010574659" watchObservedRunningTime="2025-12-06 06:55:34.565121612 +0000 UTC m=+148.019982656" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.575644 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" event={"ID":"9a24d873-f77d-43d9-a402-b7a3d9194496","Type":"ContainerStarted","Data":"f9c0b76a4d002b08f0e0c15b7b1e4b811f8c94e85cc23e4767ee705f7ac3ec1b"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.583824 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:34 crc kubenswrapper[4945]: E1206 06:55:34.585495 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:35.085470764 +0000 UTC m=+148.540331808 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.597909 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-4qjdx" podStartSLOduration=126.59788818 podStartE2EDuration="2m6.59788818s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.597405027 +0000 UTC m=+148.052266071" watchObservedRunningTime="2025-12-06 06:55:34.59788818 +0000 UTC m=+148.052749224" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.604106 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" event={"ID":"4fdae04d-27bb-4419-88cf-0ff25eeb6ff8","Type":"ContainerStarted","Data":"d8f20b1241aceb859fd3a2ba00f1ebfd5947df10ba46a58864832e0b69790de8"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.610111 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gc6qb" event={"ID":"c0b86555-314f-41f4-8a5d-5aeb356fc01f","Type":"ContainerStarted","Data":"df6da759c3e31a8c7a62da0831ad54c65ce8a3710a224a7b71c0f682fe612aea"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.649246 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" event={"ID":"957f12af-ad21-4c4a-97a2-ab1b824aa5a0","Type":"ContainerStarted","Data":"58855a5cb5145e6544f569cb20d94290c3da7d7fdae7b3429bfec3d21f8c4a95"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.650420 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.662970 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" podStartSLOduration=125.662950453 podStartE2EDuration="2m5.662950453s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.661528316 +0000 UTC m=+148.116389360" watchObservedRunningTime="2025-12-06 06:55:34.662950453 +0000 UTC m=+148.117811497" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.679582 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" event={"ID":"c9ef2f41-8ec0-48fe-8b43-a2d230b60392","Type":"ContainerStarted","Data":"adcf5d6b2d66a3dbd1f689cfdc36388ba53d9018836ecf2a13e3b8a442a209f7"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.679641 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" event={"ID":"c9ef2f41-8ec0-48fe-8b43-a2d230b60392","Type":"ContainerStarted","Data":"080ad2ed512e9cbb75a740d6e9f8165b787284d8c235b6d5af07f5dcb5024bc2"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.680389 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.692042 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-mgmz8" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.700421 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.700601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:34 crc kubenswrapper[4945]: E1206 06:55:34.701319 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:35.201297097 +0000 UTC m=+148.656158131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.708438 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.724492 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jv77f" event={"ID":"23ec3b33-0492-49bd-b1df-aa5aaac01315","Type":"ContainerStarted","Data":"066bc6cb5b529fdace0c90c1eb1da6d5bf293be09e988366052b5394bb3c65cf"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.764024 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d5rzf" podStartSLOduration=126.764006169 podStartE2EDuration="2m6.764006169s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.762854599 +0000 UTC m=+148.217715653" watchObservedRunningTime="2025-12-06 06:55:34.764006169 +0000 UTC m=+148.218867213" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.765594 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-n2zb6" podStartSLOduration=126.765586801 podStartE2EDuration="2m6.765586801s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.725568083 +0000 UTC m=+148.180429117" watchObservedRunningTime="2025-12-06 06:55:34.765586801 +0000 UTC m=+148.220447845" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.802690 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" event={"ID":"e86a5672-d1fa-4706-a073-90193128f097","Type":"ContainerStarted","Data":"3e3d56539da569345af11f9d6d4d1600297e489d214b65bf6cfe9a0cfcebd72f"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.802749 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" event={"ID":"e86a5672-d1fa-4706-a073-90193128f097","Type":"ContainerStarted","Data":"7f83545f11bca582a916f637bc7544784f1751de034abc6a0e2b0d04af5d383e"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.802797 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.803345 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.803403 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.803479 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:34 crc kubenswrapper[4945]: E1206 06:55:34.803676 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:35.303655928 +0000 UTC m=+148.758516972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.829076 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.850905 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.866683 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p8hsr" event={"ID":"3e69fe71-4b05-43dd-a6a8-dd10ec56c938","Type":"ContainerStarted","Data":"998302894bb6656f34538017f7e8e9ec6cd384c6cf01b9b101f746b389c8f307"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.878304 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.905926 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:34 crc kubenswrapper[4945]: E1206 06:55:34.907734 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:35.407717743 +0000 UTC m=+148.862578787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.915108 4945 generic.go:334] "Generic (PLEG): container finished" podID="5fc7fa8b-727b-439a-bf88-7210f1387f6d" containerID="440314421be6c6b4584dbb5d3cbe5433578463228fe7181505138322b0f14a67" exitCode=0 Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.915210 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" event={"ID":"5fc7fa8b-727b-439a-bf88-7210f1387f6d","Type":"ContainerDied","Data":"440314421be6c6b4584dbb5d3cbe5433578463228fe7181505138322b0f14a67"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.935360 4945 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-6kqp6 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.935441 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" podUID="9cf8e342-cc9f-4017-87de-ea4f1dcaa396" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.938880 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhwkm" podStartSLOduration=126.938860468 podStartE2EDuration="2m6.938860468s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.937143753 +0000 UTC m=+148.392004787" watchObservedRunningTime="2025-12-06 06:55:34.938860468 +0000 UTC m=+148.393721512" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.987599 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" podStartSLOduration=126.987576224 podStartE2EDuration="2m6.987576224s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:34.987140362 +0000 UTC m=+148.442001406" watchObservedRunningTime="2025-12-06 06:55:34.987576224 +0000 UTC m=+148.442437268" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.989979 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv" event={"ID":"5b5df8d0-7bf5-478b-adb7-eb7291da8935","Type":"ContainerStarted","Data":"0ec553a8fe2f487e74ffe56763cc79af9ccbede8aacf6ca70b894725b504f081"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.990031 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv" event={"ID":"5b5df8d0-7bf5-478b-adb7-eb7291da8935","Type":"ContainerStarted","Data":"6d0cf2894d7f71f4d5a5682aa935a8d950f191e28c281204c003ef9890956aba"} Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.990372 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:55:34 crc kubenswrapper[4945]: I1206 06:55:34.990584 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.015726 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.015844 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:35.515823494 +0000 UTC m=+148.970684538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.017039 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.022881 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:35.522859658 +0000 UTC m=+148.977720702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.065955 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.081531 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.090685 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.127175 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.128198 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:35.628151415 +0000 UTC m=+149.083012459 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.235512 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.235883 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:35.735867376 +0000 UTC m=+149.190728420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.346757 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.347326 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:35.847300504 +0000 UTC m=+149.302161548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.389009 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" podStartSLOduration=127.388986365 podStartE2EDuration="2m7.388986365s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:35.369055793 +0000 UTC m=+148.823916837" watchObservedRunningTime="2025-12-06 06:55:35.388986365 +0000 UTC m=+148.843847409" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.389897 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-csg4k"] Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.390902 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.392042 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.414531 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.416571 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xsktd"] Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.417848 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.430168 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.438636 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4blg9" podStartSLOduration=127.438614865 podStartE2EDuration="2m7.438614865s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:35.429910827 +0000 UTC m=+148.884771871" watchObservedRunningTime="2025-12-06 06:55:35.438614865 +0000 UTC m=+148.893475909" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.449741 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-utilities\") pod \"certified-operators-csg4k\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.449803 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.449872 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd96d\" (UniqueName: \"kubernetes.io/projected/fb6b366b-20cd-404c-be4c-b07df65de3bb-kube-api-access-kd96d\") pod \"community-operators-xsktd\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.449901 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-catalog-content\") pod \"community-operators-xsktd\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.449930 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-catalog-content\") pod \"certified-operators-csg4k\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.449969 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-utilities\") pod \"community-operators-xsktd\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.449995 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcmfd\" (UniqueName: \"kubernetes.io/projected/e315043e-3530-4af3-992b-8647e92bfa2d-kube-api-access-xcmfd\") pod \"certified-operators-csg4k\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.450395 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:35.950379703 +0000 UTC m=+149.405240747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.451251 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-csg4k"] Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.525912 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xsktd"] Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.552102 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.552295 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-catalog-content\") pod \"certified-operators-csg4k\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.552339 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-utilities\") pod \"community-operators-xsktd\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.552357 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcmfd\" (UniqueName: \"kubernetes.io/projected/e315043e-3530-4af3-992b-8647e92bfa2d-kube-api-access-xcmfd\") pod \"certified-operators-csg4k\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.552389 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-utilities\") pod \"certified-operators-csg4k\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.552448 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd96d\" (UniqueName: \"kubernetes.io/projected/fb6b366b-20cd-404c-be4c-b07df65de3bb-kube-api-access-kd96d\") pod \"community-operators-xsktd\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.552469 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-catalog-content\") pod \"community-operators-xsktd\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.552891 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-catalog-content\") pod \"community-operators-xsktd\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.553308 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.053290958 +0000 UTC m=+149.508152002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.553592 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-catalog-content\") pod \"certified-operators-csg4k\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.553811 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-utilities\") pod \"community-operators-xsktd\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.554297 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-utilities\") pod \"certified-operators-csg4k\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.586766 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" podStartSLOduration=126.586739384 podStartE2EDuration="2m6.586739384s" podCreationTimestamp="2025-12-06 06:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:35.582956594 +0000 UTC m=+149.037817638" watchObservedRunningTime="2025-12-06 06:55:35.586739384 +0000 UTC m=+149.041600428" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.587078 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-82ng8"] Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.603767 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.609327 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.630486 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:35 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:35 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:35 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.630557 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.659256 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.659411 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-utilities\") pod \"certified-operators-82ng8\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.659466 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-catalog-content\") pod \"certified-operators-82ng8\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.659494 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzjts\" (UniqueName: \"kubernetes.io/projected/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-kube-api-access-vzjts\") pod \"certified-operators-82ng8\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.659881 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.159864978 +0000 UTC m=+149.614726022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.691534 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd96d\" (UniqueName: \"kubernetes.io/projected/fb6b366b-20cd-404c-be4c-b07df65de3bb-kube-api-access-kd96d\") pod \"community-operators-xsktd\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.712391 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcmfd\" (UniqueName: \"kubernetes.io/projected/e315043e-3530-4af3-992b-8647e92bfa2d-kube-api-access-xcmfd\") pod \"certified-operators-csg4k\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.718657 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.733941 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-82ng8"] Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.760697 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.760897 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.260851193 +0000 UTC m=+149.715712237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.761572 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-utilities\") pod \"certified-operators-82ng8\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.761646 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-catalog-content\") pod \"certified-operators-82ng8\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.761674 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzjts\" (UniqueName: \"kubernetes.io/projected/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-kube-api-access-vzjts\") pod \"certified-operators-82ng8\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.761720 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.762070 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-utilities\") pod \"certified-operators-82ng8\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.762205 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.262186788 +0000 UTC m=+149.717047842 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.762337 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-catalog-content\") pod \"certified-operators-82ng8\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.767652 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.857615 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8zwsd"] Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.858810 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.862749 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.863153 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.363136091 +0000 UTC m=+149.817997135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.863421 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8zwsd"] Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.871925 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzjts\" (UniqueName: \"kubernetes.io/projected/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-kube-api-access-vzjts\") pod \"certified-operators-82ng8\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.964099 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-utilities\") pod \"community-operators-8zwsd\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.964176 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.964196 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-catalog-content\") pod \"community-operators-8zwsd\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:35 crc kubenswrapper[4945]: I1206 06:55:35.964226 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgmzj\" (UniqueName: \"kubernetes.io/projected/5e37b344-d572-4952-982c-8924cdf9600d-kube-api-access-fgmzj\") pod \"community-operators-8zwsd\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:35 crc kubenswrapper[4945]: E1206 06:55:35.964586 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.464572747 +0000 UTC m=+149.919433791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:35.999402 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.039255 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-p8hsr" podStartSLOduration=128.039227972 podStartE2EDuration="2m8.039227972s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:35.947988723 +0000 UTC m=+149.402849767" watchObservedRunningTime="2025-12-06 06:55:36.039227972 +0000 UTC m=+149.494089016" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.073331 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.073937 4945 generic.go:334] "Generic (PLEG): container finished" podID="9a24d873-f77d-43d9-a402-b7a3d9194496" containerID="f9c0b76a4d002b08f0e0c15b7b1e4b811f8c94e85cc23e4767ee705f7ac3ec1b" exitCode=0 Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.074049 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" event={"ID":"9a24d873-f77d-43d9-a402-b7a3d9194496","Type":"ContainerDied","Data":"f9c0b76a4d002b08f0e0c15b7b1e4b811f8c94e85cc23e4767ee705f7ac3ec1b"} Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.074065 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-catalog-content\") pod \"community-operators-8zwsd\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.074137 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgmzj\" (UniqueName: \"kubernetes.io/projected/5e37b344-d572-4952-982c-8924cdf9600d-kube-api-access-fgmzj\") pod \"community-operators-8zwsd\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.074219 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-utilities\") pod \"community-operators-8zwsd\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.074744 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-catalog-content\") pod \"community-operators-8zwsd\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.074884 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-utilities\") pod \"community-operators-8zwsd\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:36 crc kubenswrapper[4945]: E1206 06:55:36.075299 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.575254156 +0000 UTC m=+150.030115200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.117225 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ggwcv" podStartSLOduration=128.117203774 podStartE2EDuration="2m8.117203774s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:36.099432529 +0000 UTC m=+149.554293573" watchObservedRunningTime="2025-12-06 06:55:36.117203774 +0000 UTC m=+149.572064818" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.119951 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" event={"ID":"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88","Type":"ContainerStarted","Data":"57f6ba6a920d6b3b3e604b7cb87da3d78961f16ed52a7756b69a874a9f092a8a"} Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.153073 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" event={"ID":"5fc7fa8b-727b-439a-bf88-7210f1387f6d","Type":"ContainerStarted","Data":"71b22a1de1ee2d5beeacf21d5091214a8c56b12e3acb53ffa85960f383edf091"} Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.153124 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.167647 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgmzj\" (UniqueName: \"kubernetes.io/projected/5e37b344-d572-4952-982c-8924cdf9600d-kube-api-access-fgmzj\") pod \"community-operators-8zwsd\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.175775 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:36 crc kubenswrapper[4945]: E1206 06:55:36.176536 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.676521378 +0000 UTC m=+150.131382412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.235482 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" event={"ID":"fd597d15-c58a-4569-9482-70a16d2a3133","Type":"ContainerStarted","Data":"f66714e2c0b1f9742ec701d1b8349dfcb5faf4c8e6da616b7672da95c006e85d"} Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.235532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" event={"ID":"fd597d15-c58a-4569-9482-70a16d2a3133","Type":"ContainerStarted","Data":"5885cf9d6d8fbb2599e53ce7b1114ba2e186ab694da346328671caeae4cdb4f2"} Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.237700 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.279157 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:36 crc kubenswrapper[4945]: E1206 06:55:36.282475 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.78240089 +0000 UTC m=+150.237261934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.352732 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" podStartSLOduration=128.352705601 podStartE2EDuration="2m8.352705601s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:36.348977863 +0000 UTC m=+149.803838907" watchObservedRunningTime="2025-12-06 06:55:36.352705601 +0000 UTC m=+149.807566645" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.360303 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rnq6j" event={"ID":"6e0a6aac-8db4-45b0-9d8d-bbd99962bd3c","Type":"ContainerStarted","Data":"0f619d2a698aaa3fba17da40d06817397b6fe69a35903d9e098a211795ff5f41"} Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.374753 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-lpl6d container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.374847 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lpl6d" podUID="5bfd04a8-aad7-482a-9e68-f74c9cff00ba" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.375603 4945 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r8skx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.375628 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" podUID="8d7466cb-add7-436c-8641-0bf1bda1afbd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.381794 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:36 crc kubenswrapper[4945]: E1206 06:55:36.390585 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.890560102 +0000 UTC m=+150.345421146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.478958 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" podStartSLOduration=128.478936386 podStartE2EDuration="2m8.478936386s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:36.477116549 +0000 UTC m=+149.931977583" watchObservedRunningTime="2025-12-06 06:55:36.478936386 +0000 UTC m=+149.933797430" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.479058 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6kqp6" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.484254 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:36 crc kubenswrapper[4945]: E1206 06:55:36.486151 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:36.986130755 +0000 UTC m=+150.440991799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.596758 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:36 crc kubenswrapper[4945]: E1206 06:55:36.597316 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.097294106 +0000 UTC m=+150.552155150 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.612020 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:36 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:36 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:36 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.612077 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.699939 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:36 crc kubenswrapper[4945]: E1206 06:55:36.700229 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.20019363 +0000 UTC m=+150.655054664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.700827 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:36 crc kubenswrapper[4945]: E1206 06:55:36.701265 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.201238368 +0000 UTC m=+150.656099572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.802960 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:36 crc kubenswrapper[4945]: E1206 06:55:36.803401 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.303379112 +0000 UTC m=+150.758240156 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:36 crc kubenswrapper[4945]: I1206 06:55:36.904631 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:36 crc kubenswrapper[4945]: E1206 06:55:36.905080 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.405064215 +0000 UTC m=+150.859925249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.006205 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.006994 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.506958673 +0000 UTC m=+150.961819847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.007163 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.008288 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.508258517 +0000 UTC m=+150.963119561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.109213 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.109558 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.609538649 +0000 UTC m=+151.064399693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.109943 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.110453 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.610440853 +0000 UTC m=+151.065301887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.132448 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xsktd"] Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.147737 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-csg4k"] Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.214174 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5w5w9"] Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.215964 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.214296 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.214412 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.714380985 +0000 UTC m=+151.169242029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.216888 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.217501 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.717481646 +0000 UTC m=+151.172342690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.251040 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.251571 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5w5w9"] Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.323469 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.323693 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-catalog-content\") pod \"redhat-marketplace-5w5w9\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.323723 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-utilities\") pod \"redhat-marketplace-5w5w9\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.323760 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq8mz\" (UniqueName: \"kubernetes.io/projected/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-kube-api-access-bq8mz\") pod \"redhat-marketplace-5w5w9\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.323874 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.823856101 +0000 UTC m=+151.278717145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.426837 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"20eac52a23cde32de2d4578784d54fc25a1dd83f9595021faf645ca90ab1217a"} Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.427731 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-catalog-content\") pod \"redhat-marketplace-5w5w9\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.427777 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-utilities\") pod \"redhat-marketplace-5w5w9\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.427809 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.427848 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq8mz\" (UniqueName: \"kubernetes.io/projected/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-kube-api-access-bq8mz\") pod \"redhat-marketplace-5w5w9\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.428677 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-catalog-content\") pod \"redhat-marketplace-5w5w9\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.428918 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-utilities\") pod \"redhat-marketplace-5w5w9\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.429303 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:37.929273441 +0000 UTC m=+151.384134485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.445515 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsktd" event={"ID":"fb6b366b-20cd-404c-be4c-b07df65de3bb","Type":"ContainerStarted","Data":"7ffa8fe26638284cfb1811dbe235580ab403321bf83be87a250da3e7e8de6fbc"} Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.460194 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-82ng8"] Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.494842 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" event={"ID":"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88","Type":"ContainerStarted","Data":"287bbb04ebcf9949af283ab57b8b235997bc505fc7d415f788cc83318a5b52ae"} Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.494909 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" event={"ID":"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88","Type":"ContainerStarted","Data":"8926aa545cd35e6e3a6ea6c51d3b9a456af72fd50d58671afe530895267b66fd"} Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.501724 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq8mz\" (UniqueName: \"kubernetes.io/projected/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-kube-api-access-bq8mz\") pod \"redhat-marketplace-5w5w9\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.509540 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csg4k" event={"ID":"e315043e-3530-4af3-992b-8647e92bfa2d","Type":"ContainerStarted","Data":"3643b4b1d32edf9c9db7bd456deba388815b4d99f427f9e11e074037d7163987"} Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.523710 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b4b00687bd12fefc5a15fb62b284422aedc6cf371e77bc9ca4d46a44a89c17a5"} Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.530955 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.531367 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.031346594 +0000 UTC m=+151.486207638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.531767 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8zwsd"] Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.545838 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c28cabf6433a878783e0a1fb4922b651bdbab7a3542a846a3a754b3ab501f9d3"} Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.545887 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3f9a3bf7492c559f04392dc69e43ffaf6da980a2a872434a6a83acfaf5e0e636"} Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.547213 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-lpl6d container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.547257 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lpl6d" podUID="5bfd04a8-aad7-482a-9e68-f74c9cff00ba" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.569455 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.586880 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.609709 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:37 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:37 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:37 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.610151 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.639104 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.643599 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.143575583 +0000 UTC m=+151.598436807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.740191 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.740614 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.240591023 +0000 UTC m=+151.695452067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.793895 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.808035 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.817028 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.817687 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.818817 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.843346 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33a01e5b-677b-4c61-90bc-a65d00f2302a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"33a01e5b-677b-4c61-90bc-a65d00f2302a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.843450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/33a01e5b-677b-4c61-90bc-a65d00f2302a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"33a01e5b-677b-4c61-90bc-a65d00f2302a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.843518 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.843857 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.343843737 +0000 UTC m=+151.798704781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.944650 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.944983 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/33a01e5b-677b-4c61-90bc-a65d00f2302a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"33a01e5b-677b-4c61-90bc-a65d00f2302a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.945063 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33a01e5b-677b-4c61-90bc-a65d00f2302a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"33a01e5b-677b-4c61-90bc-a65d00f2302a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 06:55:37 crc kubenswrapper[4945]: E1206 06:55:37.945610 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.445591511 +0000 UTC m=+151.900452555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.945647 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/33a01e5b-677b-4c61-90bc-a65d00f2302a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"33a01e5b-677b-4c61-90bc-a65d00f2302a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.968877 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33a01e5b-677b-4c61-90bc-a65d00f2302a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"33a01e5b-677b-4c61-90bc-a65d00f2302a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 06:55:37 crc kubenswrapper[4945]: I1206 06:55:37.997876 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5w5w9"] Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.046490 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.047052 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.547029837 +0000 UTC m=+152.001890881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.147897 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.148113 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.648070933 +0000 UTC m=+152.102932007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.148250 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.148739 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.64872567 +0000 UTC m=+152.103586714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.159630 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.188452 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.188539 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.198512 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.221555 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.221619 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.249831 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.251053 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.751025569 +0000 UTC m=+152.205886613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.353184 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.353888 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.853860732 +0000 UTC m=+152.308721816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.373883 4945 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 06 06:55:38 crc kubenswrapper[4945]: W1206 06:55:38.393431 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c50068a_69e4_4063_8f08_e4dc4e51fc8d.slice/crio-071f22cbc7765303f63617f236b07b4ddb13c6b8f96fd788658a0670f2c22992 WatchSource:0}: Error finding container 071f22cbc7765303f63617f236b07b4ddb13c6b8f96fd788658a0670f2c22992: Status 404 returned error can't find the container with id 071f22cbc7765303f63617f236b07b4ddb13c6b8f96fd788658a0670f2c22992 Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.427457 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rdjtn"] Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.428993 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.444747 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.444879 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fvjzs"] Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.445217 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a24d873-f77d-43d9-a402-b7a3d9194496" containerName="collect-profiles" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.445232 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a24d873-f77d-43d9-a402-b7a3d9194496" containerName="collect-profiles" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.445381 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a24d873-f77d-43d9-a402-b7a3d9194496" containerName="collect-profiles" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.446377 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.449381 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.461306 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.461502 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-catalog-content\") pod \"redhat-marketplace-rdjtn\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.461587 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-utilities\") pod \"redhat-marketplace-rdjtn\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.461652 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h8vw\" (UniqueName: \"kubernetes.io/projected/5d88a868-4630-4744-b0ae-b73f8bdf7331-kube-api-access-8h8vw\") pod \"redhat-marketplace-rdjtn\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.461683 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdjtn"] Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.462016 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:38.961988403 +0000 UTC m=+152.416849447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.462153 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.487636 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zxzv5" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.514067 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fvjzs"] Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.562070 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a24d873-f77d-43d9-a402-b7a3d9194496-config-volume\") pod \"9a24d873-f77d-43d9-a402-b7a3d9194496\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.562832 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkfc5\" (UniqueName: \"kubernetes.io/projected/9a24d873-f77d-43d9-a402-b7a3d9194496-kube-api-access-bkfc5\") pod \"9a24d873-f77d-43d9-a402-b7a3d9194496\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.563006 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a24d873-f77d-43d9-a402-b7a3d9194496-secret-volume\") pod \"9a24d873-f77d-43d9-a402-b7a3d9194496\" (UID: \"9a24d873-f77d-43d9-a402-b7a3d9194496\") " Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.566485 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a24d873-f77d-43d9-a402-b7a3d9194496-config-volume" (OuterVolumeSpecName: "config-volume") pod "9a24d873-f77d-43d9-a402-b7a3d9194496" (UID: "9a24d873-f77d-43d9-a402-b7a3d9194496"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.568238 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-catalog-content\") pod \"redhat-operators-fvjzs\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.570363 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-utilities\") pod \"redhat-marketplace-rdjtn\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.570551 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfw44\" (UniqueName: \"kubernetes.io/projected/319506ca-cc5d-424b-a84f-e7fab30c8a76-kube-api-access-kfw44\") pod \"redhat-operators-fvjzs\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.570702 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.570831 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-utilities\") pod \"redhat-operators-fvjzs\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.570984 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h8vw\" (UniqueName: \"kubernetes.io/projected/5d88a868-4630-4744-b0ae-b73f8bdf7331-kube-api-access-8h8vw\") pod \"redhat-marketplace-rdjtn\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.571115 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-catalog-content\") pod \"redhat-marketplace-rdjtn\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.571241 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a24d873-f77d-43d9-a402-b7a3d9194496-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.573736 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:39.073719229 +0000 UTC m=+152.528580273 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.574468 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-utilities\") pod \"redhat-marketplace-rdjtn\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.576025 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-catalog-content\") pod \"redhat-marketplace-rdjtn\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.577110 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a24d873-f77d-43d9-a402-b7a3d9194496-kube-api-access-bkfc5" (OuterVolumeSpecName: "kube-api-access-bkfc5") pod "9a24d873-f77d-43d9-a402-b7a3d9194496" (UID: "9a24d873-f77d-43d9-a402-b7a3d9194496"). InnerVolumeSpecName "kube-api-access-bkfc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.579044 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a24d873-f77d-43d9-a402-b7a3d9194496-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9a24d873-f77d-43d9-a402-b7a3d9194496" (UID: "9a24d873-f77d-43d9-a402-b7a3d9194496"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.588054 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lnz2w"] Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.589258 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.604450 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.608107 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lnz2w"] Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.614628 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h8vw\" (UniqueName: \"kubernetes.io/projected/5d88a868-4630-4744-b0ae-b73f8bdf7331-kube-api-access-8h8vw\") pod \"redhat-marketplace-rdjtn\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.619082 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:38 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:38 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:38 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.619169 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.648865 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2572326cdd548e268381f31e3fccb8e3224fe3ff4018e831a63a63f0f18573a7"} Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.683008 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zwsd" event={"ID":"5e37b344-d572-4952-982c-8924cdf9600d","Type":"ContainerStarted","Data":"f62a1839e5b093efc9c68f6b10260e6572723f04e1d4e32ec1713c4fc70fccc9"} Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.689045 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82ng8" event={"ID":"645fe9fd-1822-4430-8a4c-17f9eeb34e4e","Type":"ContainerStarted","Data":"7f8f8c1b1c00f5b742aa19e066979f4fb9c509d9ebbbe7afbb70a4285acda9bb"} Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.689087 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82ng8" event={"ID":"645fe9fd-1822-4430-8a4c-17f9eeb34e4e","Type":"ContainerStarted","Data":"22ddf88a4070c220690bf412892c856d3c5b8ad7289c53d9ec181bd59823cd0f"} Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.697630 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.698152 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:39.198111877 +0000 UTC m=+152.652972921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.700858 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-utilities\") pod \"redhat-operators-lnz2w\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.701340 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-catalog-content\") pod \"redhat-operators-lnz2w\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.701624 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-catalog-content\") pod \"redhat-operators-fvjzs\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.702026 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l5gp\" (UniqueName: \"kubernetes.io/projected/a82181e2-5f33-410a-bec7-99ff14853494-kube-api-access-4l5gp\") pod \"redhat-operators-lnz2w\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.702766 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfw44\" (UniqueName: \"kubernetes.io/projected/319506ca-cc5d-424b-a84f-e7fab30c8a76-kube-api-access-kfw44\") pod \"redhat-operators-fvjzs\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.704729 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.705130 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-utilities\") pod \"redhat-operators-fvjzs\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.706422 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkfc5\" (UniqueName: \"kubernetes.io/projected/9a24d873-f77d-43d9-a402-b7a3d9194496-kube-api-access-bkfc5\") on node \"crc\" DevicePath \"\"" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.706446 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a24d873-f77d-43d9-a402-b7a3d9194496-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.703562 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-catalog-content\") pod \"redhat-operators-fvjzs\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.702842 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.706993 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:39.206958258 +0000 UTC m=+152.661819302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.709296 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"33a01e5b-677b-4c61-90bc-a65d00f2302a","Type":"ContainerStarted","Data":"94c0586a2b413e7480f733486884e22b923a004f52d5aaf0b37f9c1bc9fba2a9"} Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.710345 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-utilities\") pod \"redhat-operators-fvjzs\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.728230 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfw44\" (UniqueName: \"kubernetes.io/projected/319506ca-cc5d-424b-a84f-e7fab30c8a76-kube-api-access-kfw44\") pod \"redhat-operators-fvjzs\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.733200 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csg4k" event={"ID":"e315043e-3530-4af3-992b-8647e92bfa2d","Type":"ContainerStarted","Data":"00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd"} Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.746825 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w5w9" event={"ID":"4c50068a-69e4-4063-8f08-e4dc4e51fc8d","Type":"ContainerStarted","Data":"071f22cbc7765303f63617f236b07b4ddb13c6b8f96fd788658a0670f2c22992"} Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.757769 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"94b7741f06cdd9e6085d5c86f892ecc45143fa731d368ef72a7e5c0cd72eb163"} Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.758441 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.761208 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.763802 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf" event={"ID":"9a24d873-f77d-43d9-a402-b7a3d9194496","Type":"ContainerDied","Data":"48dc03749c61be8e42d4b7c2baecd2ec9053d407d1e6ca1b38e3d12e861ef6db"} Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.763830 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48dc03749c61be8e42d4b7c2baecd2ec9053d407d1e6ca1b38e3d12e861ef6db" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.812960 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.813024 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.813604 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d2p6r" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.814019 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.838021 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.849645 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.850412 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:39.350378534 +0000 UTC m=+152.805239578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.856449 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-utilities\") pod \"redhat-operators-lnz2w\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.856749 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-catalog-content\") pod \"redhat-operators-lnz2w\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.857027 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l5gp\" (UniqueName: \"kubernetes.io/projected/a82181e2-5f33-410a-bec7-99ff14853494-kube-api-access-4l5gp\") pod \"redhat-operators-lnz2w\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.858140 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.860153 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-utilities\") pod \"redhat-operators-lnz2w\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.861806 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-catalog-content\") pod \"redhat-operators-lnz2w\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.865858 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:39.365825098 +0000 UTC m=+152.820686142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.932495 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l5gp\" (UniqueName: \"kubernetes.io/projected/a82181e2-5f33-410a-bec7-99ff14853494-kube-api-access-4l5gp\") pod \"redhat-operators-lnz2w\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:38 crc kubenswrapper[4945]: I1206 06:55:38.964074 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:38 crc kubenswrapper[4945]: E1206 06:55:38.965341 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 06:55:39.465320634 +0000 UTC m=+152.920181678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.047844 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-lpl6d container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.047913 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lpl6d" podUID="5bfd04a8-aad7-482a-9e68-f74c9cff00ba" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.048153 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-lpl6d container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.048217 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lpl6d" podUID="5bfd04a8-aad7-482a-9e68-f74c9cff00ba" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.049559 4945 patch_prober.go:28] interesting pod/console-f9d7485db-j2tjw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.049590 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-j2tjw" podUID="50e16132-af33-4dbf-b4f4-a9b0c8554e24" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.066791 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:39 crc kubenswrapper[4945]: E1206 06:55:39.067317 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 06:55:39.567264233 +0000 UTC m=+153.022125277 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8ppq" (UID: "401f27c7-5383-4ade-839b-59a92b137752") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.139150 4945 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-06T06:55:38.374495432Z","Handler":null,"Name":""} Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.164407 4945 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.164455 4945 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.167969 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.191761 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.604166 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:39 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:39 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:39 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.604595 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.658001 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.665764 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.666729 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.666769 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.666812 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fvjzs"] Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.666832 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdjtn"] Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.674263 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.685569 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.685621 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.791595 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdjtn" event={"ID":"5d88a868-4630-4744-b0ae-b73f8bdf7331","Type":"ContainerStarted","Data":"605db9afa1aa66d7345821180794d0a566bafc1e5a1edcd919d02a33ee0651e8"} Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.801126 4945 generic.go:334] "Generic (PLEG): container finished" podID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerID="bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914" exitCode=0 Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.801256 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsktd" event={"ID":"fb6b366b-20cd-404c-be4c-b07df65de3bb","Type":"ContainerDied","Data":"bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914"} Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.822583 4945 patch_prober.go:28] interesting pod/apiserver-76f77b778f-ktrcx container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 06 06:55:39 crc kubenswrapper[4945]: [+]log ok Dec 06 06:55:39 crc kubenswrapper[4945]: [+]etcd ok Dec 06 06:55:39 crc kubenswrapper[4945]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 06 06:55:39 crc kubenswrapper[4945]: [+]poststarthook/generic-apiserver-start-informers ok Dec 06 06:55:39 crc kubenswrapper[4945]: [+]poststarthook/max-in-flight-filter ok Dec 06 06:55:39 crc kubenswrapper[4945]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 06 06:55:39 crc kubenswrapper[4945]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 06 06:55:39 crc kubenswrapper[4945]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 06 06:55:39 crc kubenswrapper[4945]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 06 06:55:39 crc kubenswrapper[4945]: [+]poststarthook/project.openshift.io-projectcache ok Dec 06 06:55:39 crc kubenswrapper[4945]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 06 06:55:39 crc kubenswrapper[4945]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Dec 06 06:55:39 crc kubenswrapper[4945]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 06 06:55:39 crc kubenswrapper[4945]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 06 06:55:39 crc kubenswrapper[4945]: livez check failed Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.822648 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" podUID="fd597d15-c58a-4569-9482-70a16d2a3133" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.826788 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" event={"ID":"e9bf5fde-fb1b-4e2f-ab6f-97aef4fa9c88","Type":"ContainerStarted","Data":"2e54296dbc8b6d32853bd3dff490374022817f662145bc5a194b33fe184343f1"} Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.835980 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8ppq\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.844000 4945 generic.go:334] "Generic (PLEG): container finished" podID="e315043e-3530-4af3-992b-8647e92bfa2d" containerID="00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd" exitCode=0 Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.844081 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csg4k" event={"ID":"e315043e-3530-4af3-992b-8647e92bfa2d","Type":"ContainerDied","Data":"00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd"} Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.846942 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvjzs" event={"ID":"319506ca-cc5d-424b-a84f-e7fab30c8a76","Type":"ContainerStarted","Data":"dc4897c7f8eb297444fa4f2bf2865d99788fc72346bbff9f3cfd2c84ecdb337e"} Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.847953 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e37b344-d572-4952-982c-8924cdf9600d" containerID="a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198" exitCode=0 Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.848011 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zwsd" event={"ID":"5e37b344-d572-4952-982c-8924cdf9600d","Type":"ContainerDied","Data":"a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198"} Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.849880 4945 generic.go:334] "Generic (PLEG): container finished" podID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerID="7f8f8c1b1c00f5b742aa19e066979f4fb9c509d9ebbbe7afbb70a4285acda9bb" exitCode=0 Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.849959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82ng8" event={"ID":"645fe9fd-1822-4430-8a4c-17f9eeb34e4e","Type":"ContainerDied","Data":"7f8f8c1b1c00f5b742aa19e066979f4fb9c509d9ebbbe7afbb70a4285acda9bb"} Dec 06 06:55:39 crc kubenswrapper[4945]: I1206 06:55:39.859021 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-cf4n5" podStartSLOduration=13.858987975 podStartE2EDuration="13.858987975s" podCreationTimestamp="2025-12-06 06:55:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:39.855810632 +0000 UTC m=+153.310671686" watchObservedRunningTime="2025-12-06 06:55:39.858987975 +0000 UTC m=+153.313849019" Dec 06 06:55:40 crc kubenswrapper[4945]: I1206 06:55:40.012978 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:40 crc kubenswrapper[4945]: I1206 06:55:40.046243 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lnz2w"] Dec 06 06:55:40 crc kubenswrapper[4945]: W1206 06:55:40.067793 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda82181e2_5f33_410a_bec7_99ff14853494.slice/crio-e3b27405e990eef62baa555435c6531ae0b2e71158e9093b9c055a63c0c116c2 WatchSource:0}: Error finding container e3b27405e990eef62baa555435c6531ae0b2e71158e9093b9c055a63c0c116c2: Status 404 returned error can't find the container with id e3b27405e990eef62baa555435c6531ae0b2e71158e9093b9c055a63c0c116c2 Dec 06 06:55:40 crc kubenswrapper[4945]: I1206 06:55:40.234609 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8ppq"] Dec 06 06:55:40 crc kubenswrapper[4945]: W1206 06:55:40.241948 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod401f27c7_5383_4ade_839b_59a92b137752.slice/crio-1a4f727d0af718df8e80387b21d10b276779c17a4e78ccaa887430fe783ede59 WatchSource:0}: Error finding container 1a4f727d0af718df8e80387b21d10b276779c17a4e78ccaa887430fe783ede59: Status 404 returned error can't find the container with id 1a4f727d0af718df8e80387b21d10b276779c17a4e78ccaa887430fe783ede59 Dec 06 06:55:40 crc kubenswrapper[4945]: I1206 06:55:40.604720 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:40 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:40 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:40 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:40 crc kubenswrapper[4945]: I1206 06:55:40.604801 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:40 crc kubenswrapper[4945]: I1206 06:55:40.858521 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" event={"ID":"401f27c7-5383-4ade-839b-59a92b137752","Type":"ContainerStarted","Data":"1a4f727d0af718df8e80387b21d10b276779c17a4e78ccaa887430fe783ede59"} Dec 06 06:55:40 crc kubenswrapper[4945]: I1206 06:55:40.860332 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lnz2w" event={"ID":"a82181e2-5f33-410a-bec7-99ff14853494","Type":"ContainerStarted","Data":"e3b27405e990eef62baa555435c6531ae0b2e71158e9093b9c055a63c0c116c2"} Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.548396 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.550198 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.553351 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.553358 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.556623 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.605515 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:41 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:41 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:41 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.605590 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.611510 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/293101a4-3640-4d8c-9043-fe31868e746d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"293101a4-3640-4d8c-9043-fe31868e746d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.611638 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/293101a4-3640-4d8c-9043-fe31868e746d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"293101a4-3640-4d8c-9043-fe31868e746d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.713381 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/293101a4-3640-4d8c-9043-fe31868e746d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"293101a4-3640-4d8c-9043-fe31868e746d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.713527 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/293101a4-3640-4d8c-9043-fe31868e746d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"293101a4-3640-4d8c-9043-fe31868e746d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.713651 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/293101a4-3640-4d8c-9043-fe31868e746d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"293101a4-3640-4d8c-9043-fe31868e746d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.740995 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/293101a4-3640-4d8c-9043-fe31868e746d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"293101a4-3640-4d8c-9043-fe31868e746d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.873693 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" event={"ID":"401f27c7-5383-4ade-839b-59a92b137752","Type":"ContainerStarted","Data":"a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4"} Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.874502 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.879235 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.884537 4945 generic.go:334] "Generic (PLEG): container finished" podID="a82181e2-5f33-410a-bec7-99ff14853494" containerID="cfa6c17f0d891f3171a04a79c8106f794990b88e381a516d227c1a0da4060ba8" exitCode=0 Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.884620 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lnz2w" event={"ID":"a82181e2-5f33-410a-bec7-99ff14853494","Type":"ContainerDied","Data":"cfa6c17f0d891f3171a04a79c8106f794990b88e381a516d227c1a0da4060ba8"} Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.888612 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"33a01e5b-677b-4c61-90bc-a65d00f2302a","Type":"ContainerStarted","Data":"0bdfbc2d0f5a70db98a41ed4b2debc0833d0e1869bb79a3864d269cd20b039dd"} Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.902611 4945 generic.go:334] "Generic (PLEG): container finished" podID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerID="534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd" exitCode=0 Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.902820 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w5w9" event={"ID":"4c50068a-69e4-4063-8f08-e4dc4e51fc8d","Type":"ContainerDied","Data":"534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd"} Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.912660 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" podStartSLOduration=133.912637551 podStartE2EDuration="2m13.912637551s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:41.906304425 +0000 UTC m=+155.361165489" watchObservedRunningTime="2025-12-06 06:55:41.912637551 +0000 UTC m=+155.367498595" Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.915071 4945 generic.go:334] "Generic (PLEG): container finished" podID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerID="cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb" exitCode=0 Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.915271 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvjzs" event={"ID":"319506ca-cc5d-424b-a84f-e7fab30c8a76","Type":"ContainerDied","Data":"cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb"} Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.926967 4945 generic.go:334] "Generic (PLEG): container finished" podID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerID="38295fbf9d19f23c736d583ec5a9bd6d5350c0f352da8bcc6c70e8db76fedacd" exitCode=0 Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.927028 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdjtn" event={"ID":"5d88a868-4630-4744-b0ae-b73f8bdf7331","Type":"ContainerDied","Data":"38295fbf9d19f23c736d583ec5a9bd6d5350c0f352da8bcc6c70e8db76fedacd"} Dec 06 06:55:41 crc kubenswrapper[4945]: I1206 06:55:41.958301 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.958203164 podStartE2EDuration="4.958203164s" podCreationTimestamp="2025-12-06 06:55:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:41.93513341 +0000 UTC m=+155.389994464" watchObservedRunningTime="2025-12-06 06:55:41.958203164 +0000 UTC m=+155.413064208" Dec 06 06:55:42 crc kubenswrapper[4945]: I1206 06:55:42.180330 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 06:55:42 crc kubenswrapper[4945]: I1206 06:55:42.611162 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:42 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:42 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:42 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:42 crc kubenswrapper[4945]: I1206 06:55:42.611760 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:42 crc kubenswrapper[4945]: I1206 06:55:42.938378 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"293101a4-3640-4d8c-9043-fe31868e746d","Type":"ContainerStarted","Data":"43aebb28ea208b74449dbf069cb4e39f19622bb200dd42dcd982dd17e691c18e"} Dec 06 06:55:42 crc kubenswrapper[4945]: I1206 06:55:42.938846 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"293101a4-3640-4d8c-9043-fe31868e746d","Type":"ContainerStarted","Data":"b6b8b120c6e91bf6a24ac07221ee74c2501e316235821ed2afaf2b06c1065e90"} Dec 06 06:55:42 crc kubenswrapper[4945]: I1206 06:55:42.940833 4945 generic.go:334] "Generic (PLEG): container finished" podID="33a01e5b-677b-4c61-90bc-a65d00f2302a" containerID="0bdfbc2d0f5a70db98a41ed4b2debc0833d0e1869bb79a3864d269cd20b039dd" exitCode=0 Dec 06 06:55:42 crc kubenswrapper[4945]: I1206 06:55:42.942622 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"33a01e5b-677b-4c61-90bc-a65d00f2302a","Type":"ContainerDied","Data":"0bdfbc2d0f5a70db98a41ed4b2debc0833d0e1869bb79a3864d269cd20b039dd"} Dec 06 06:55:42 crc kubenswrapper[4945]: I1206 06:55:42.964534 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=1.964504185 podStartE2EDuration="1.964504185s" podCreationTimestamp="2025-12-06 06:55:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:55:42.953745423 +0000 UTC m=+156.408606467" watchObservedRunningTime="2025-12-06 06:55:42.964504185 +0000 UTC m=+156.419365229" Dec 06 06:55:43 crc kubenswrapper[4945]: I1206 06:55:43.227264 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:43 crc kubenswrapper[4945]: I1206 06:55:43.244656 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-ktrcx" Dec 06 06:55:43 crc kubenswrapper[4945]: I1206 06:55:43.357984 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 06:55:43 crc kubenswrapper[4945]: I1206 06:55:43.605007 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:43 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:43 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:43 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:43 crc kubenswrapper[4945]: I1206 06:55:43.605071 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.327158 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.495306 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33a01e5b-677b-4c61-90bc-a65d00f2302a-kube-api-access\") pod \"33a01e5b-677b-4c61-90bc-a65d00f2302a\" (UID: \"33a01e5b-677b-4c61-90bc-a65d00f2302a\") " Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.495441 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/33a01e5b-677b-4c61-90bc-a65d00f2302a-kubelet-dir\") pod \"33a01e5b-677b-4c61-90bc-a65d00f2302a\" (UID: \"33a01e5b-677b-4c61-90bc-a65d00f2302a\") " Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.495834 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/33a01e5b-677b-4c61-90bc-a65d00f2302a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "33a01e5b-677b-4c61-90bc-a65d00f2302a" (UID: "33a01e5b-677b-4c61-90bc-a65d00f2302a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.506807 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33a01e5b-677b-4c61-90bc-a65d00f2302a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "33a01e5b-677b-4c61-90bc-a65d00f2302a" (UID: "33a01e5b-677b-4c61-90bc-a65d00f2302a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.523677 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rnq6j" Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.597488 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33a01e5b-677b-4c61-90bc-a65d00f2302a-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.597526 4945 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/33a01e5b-677b-4c61-90bc-a65d00f2302a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.609084 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:44 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:44 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:44 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.609177 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.977564 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 06:55:44 crc kubenswrapper[4945]: I1206 06:55:44.980825 4945 generic.go:334] "Generic (PLEG): container finished" podID="293101a4-3640-4d8c-9043-fe31868e746d" containerID="43aebb28ea208b74449dbf069cb4e39f19622bb200dd42dcd982dd17e691c18e" exitCode=0 Dec 06 06:55:45 crc kubenswrapper[4945]: I1206 06:55:45.011351 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"33a01e5b-677b-4c61-90bc-a65d00f2302a","Type":"ContainerDied","Data":"94c0586a2b413e7480f733486884e22b923a004f52d5aaf0b37f9c1bc9fba2a9"} Dec 06 06:55:45 crc kubenswrapper[4945]: I1206 06:55:45.011398 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94c0586a2b413e7480f733486884e22b923a004f52d5aaf0b37f9c1bc9fba2a9" Dec 06 06:55:45 crc kubenswrapper[4945]: I1206 06:55:45.011433 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"293101a4-3640-4d8c-9043-fe31868e746d","Type":"ContainerDied","Data":"43aebb28ea208b74449dbf069cb4e39f19622bb200dd42dcd982dd17e691c18e"} Dec 06 06:55:45 crc kubenswrapper[4945]: I1206 06:55:45.603933 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:45 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:45 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:45 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:45 crc kubenswrapper[4945]: I1206 06:55:45.604025 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:46 crc kubenswrapper[4945]: I1206 06:55:46.296225 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 06:55:46 crc kubenswrapper[4945]: I1206 06:55:46.432184 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/293101a4-3640-4d8c-9043-fe31868e746d-kubelet-dir\") pod \"293101a4-3640-4d8c-9043-fe31868e746d\" (UID: \"293101a4-3640-4d8c-9043-fe31868e746d\") " Dec 06 06:55:46 crc kubenswrapper[4945]: I1206 06:55:46.432251 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/293101a4-3640-4d8c-9043-fe31868e746d-kube-api-access\") pod \"293101a4-3640-4d8c-9043-fe31868e746d\" (UID: \"293101a4-3640-4d8c-9043-fe31868e746d\") " Dec 06 06:55:46 crc kubenswrapper[4945]: I1206 06:55:46.432921 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/293101a4-3640-4d8c-9043-fe31868e746d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "293101a4-3640-4d8c-9043-fe31868e746d" (UID: "293101a4-3640-4d8c-9043-fe31868e746d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:55:46 crc kubenswrapper[4945]: I1206 06:55:46.442588 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/293101a4-3640-4d8c-9043-fe31868e746d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "293101a4-3640-4d8c-9043-fe31868e746d" (UID: "293101a4-3640-4d8c-9043-fe31868e746d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:55:46 crc kubenswrapper[4945]: I1206 06:55:46.534359 4945 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/293101a4-3640-4d8c-9043-fe31868e746d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 06:55:46 crc kubenswrapper[4945]: I1206 06:55:46.534474 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/293101a4-3640-4d8c-9043-fe31868e746d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 06:55:46 crc kubenswrapper[4945]: I1206 06:55:46.603353 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:46 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:46 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:46 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:46 crc kubenswrapper[4945]: I1206 06:55:46.603430 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:47 crc kubenswrapper[4945]: I1206 06:55:47.049058 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"293101a4-3640-4d8c-9043-fe31868e746d","Type":"ContainerDied","Data":"b6b8b120c6e91bf6a24ac07221ee74c2501e316235821ed2afaf2b06c1065e90"} Dec 06 06:55:47 crc kubenswrapper[4945]: I1206 06:55:47.049116 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6b8b120c6e91bf6a24ac07221ee74c2501e316235821ed2afaf2b06c1065e90" Dec 06 06:55:47 crc kubenswrapper[4945]: I1206 06:55:47.049242 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 06:55:48 crc kubenswrapper[4945]: I1206 06:55:48.394567 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:48 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:48 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:48 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:48 crc kubenswrapper[4945]: I1206 06:55:48.395396 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:48 crc kubenswrapper[4945]: I1206 06:55:48.604901 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:48 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:48 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:48 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:48 crc kubenswrapper[4945]: I1206 06:55:48.606057 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:49 crc kubenswrapper[4945]: I1206 06:55:49.072662 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-lpl6d container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 06 06:55:49 crc kubenswrapper[4945]: I1206 06:55:49.072662 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-lpl6d container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 06 06:55:49 crc kubenswrapper[4945]: I1206 06:55:49.072710 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lpl6d" podUID="5bfd04a8-aad7-482a-9e68-f74c9cff00ba" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 06 06:55:49 crc kubenswrapper[4945]: I1206 06:55:49.072756 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lpl6d" podUID="5bfd04a8-aad7-482a-9e68-f74c9cff00ba" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 06 06:55:49 crc kubenswrapper[4945]: I1206 06:55:49.073173 4945 patch_prober.go:28] interesting pod/console-f9d7485db-j2tjw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 06 06:55:49 crc kubenswrapper[4945]: I1206 06:55:49.073197 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-j2tjw" podUID="50e16132-af33-4dbf-b4f4-a9b0c8554e24" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 06 06:55:49 crc kubenswrapper[4945]: I1206 06:55:49.603913 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:49 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:49 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:49 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:49 crc kubenswrapper[4945]: I1206 06:55:49.603984 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:50 crc kubenswrapper[4945]: I1206 06:55:50.603343 4945 patch_prober.go:28] interesting pod/router-default-5444994796-p8hsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 06:55:50 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Dec 06 06:55:50 crc kubenswrapper[4945]: [+]process-running ok Dec 06 06:55:50 crc kubenswrapper[4945]: healthz check failed Dec 06 06:55:50 crc kubenswrapper[4945]: I1206 06:55:50.603404 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p8hsr" podUID="3e69fe71-4b05-43dd-a6a8-dd10ec56c938" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 06:55:50 crc kubenswrapper[4945]: I1206 06:55:50.702140 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:50 crc kubenswrapper[4945]: I1206 06:55:50.711500 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4326ced-ae97-4301-854b-222062d4764e-metrics-certs\") pod \"network-metrics-daemon-2xkp9\" (UID: \"d4326ced-ae97-4301-854b-222062d4764e\") " pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:50 crc kubenswrapper[4945]: I1206 06:55:50.867069 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2xkp9" Dec 06 06:55:51 crc kubenswrapper[4945]: I1206 06:55:51.338909 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2xkp9"] Dec 06 06:55:51 crc kubenswrapper[4945]: W1206 06:55:51.359419 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4326ced_ae97_4301_854b_222062d4764e.slice/crio-c3d9ed493cac809d9f62c81b2d9aa1f6ba52b29033a13208310ffb2754680063 WatchSource:0}: Error finding container c3d9ed493cac809d9f62c81b2d9aa1f6ba52b29033a13208310ffb2754680063: Status 404 returned error can't find the container with id c3d9ed493cac809d9f62c81b2d9aa1f6ba52b29033a13208310ffb2754680063 Dec 06 06:55:51 crc kubenswrapper[4945]: I1206 06:55:51.414394 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" event={"ID":"d4326ced-ae97-4301-854b-222062d4764e","Type":"ContainerStarted","Data":"c3d9ed493cac809d9f62c81b2d9aa1f6ba52b29033a13208310ffb2754680063"} Dec 06 06:55:51 crc kubenswrapper[4945]: I1206 06:55:51.604787 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:51 crc kubenswrapper[4945]: I1206 06:55:51.610620 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-p8hsr" Dec 06 06:55:52 crc kubenswrapper[4945]: I1206 06:55:52.432087 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" event={"ID":"d4326ced-ae97-4301-854b-222062d4764e","Type":"ContainerStarted","Data":"eb616c74dd4a48e0f2f85131a2fc95a9d514d25b72b68a10f90eccbc2c3ee973"} Dec 06 06:55:59 crc kubenswrapper[4945]: I1206 06:55:59.056911 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:55:59 crc kubenswrapper[4945]: I1206 06:55:59.058151 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-lpl6d" Dec 06 06:55:59 crc kubenswrapper[4945]: I1206 06:55:59.060906 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 06:56:00 crc kubenswrapper[4945]: I1206 06:56:00.018704 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 06:56:08 crc kubenswrapper[4945]: I1206 06:56:08.795791 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:56:08 crc kubenswrapper[4945]: I1206 06:56:08.797656 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:56:09 crc kubenswrapper[4945]: I1206 06:56:09.739448 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-62dkz" Dec 06 06:56:15 crc kubenswrapper[4945]: I1206 06:56:15.637869 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.542861 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 06:56:16 crc kubenswrapper[4945]: E1206 06:56:16.543341 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a01e5b-677b-4c61-90bc-a65d00f2302a" containerName="pruner" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.543358 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a01e5b-677b-4c61-90bc-a65d00f2302a" containerName="pruner" Dec 06 06:56:16 crc kubenswrapper[4945]: E1206 06:56:16.543370 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="293101a4-3640-4d8c-9043-fe31868e746d" containerName="pruner" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.543376 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="293101a4-3640-4d8c-9043-fe31868e746d" containerName="pruner" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.543471 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a01e5b-677b-4c61-90bc-a65d00f2302a" containerName="pruner" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.543481 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="293101a4-3640-4d8c-9043-fe31868e746d" containerName="pruner" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.543831 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.546926 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.548010 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.557405 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.698566 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"99fa4df9-99a0-4cbf-9562-15bac87fa8ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.698676 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"99fa4df9-99a0-4cbf-9562-15bac87fa8ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.800529 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"99fa4df9-99a0-4cbf-9562-15bac87fa8ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.800632 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"99fa4df9-99a0-4cbf-9562-15bac87fa8ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.800742 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"99fa4df9-99a0-4cbf-9562-15bac87fa8ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.827006 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"99fa4df9-99a0-4cbf-9562-15bac87fa8ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 06:56:16 crc kubenswrapper[4945]: I1206 06:56:16.863323 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 06:56:20 crc kubenswrapper[4945]: I1206 06:56:20.939541 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 06:56:20 crc kubenswrapper[4945]: I1206 06:56:20.941186 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:20 crc kubenswrapper[4945]: I1206 06:56:20.952017 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 06:56:20 crc kubenswrapper[4945]: I1206 06:56:20.977134 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0af9496b-2588-4cb7-9bc2-922c48b01af2-kube-api-access\") pod \"installer-9-crc\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:20 crc kubenswrapper[4945]: I1206 06:56:20.977188 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:20 crc kubenswrapper[4945]: I1206 06:56:20.977240 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-var-lock\") pod \"installer-9-crc\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:21 crc kubenswrapper[4945]: I1206 06:56:21.078653 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-var-lock\") pod \"installer-9-crc\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:21 crc kubenswrapper[4945]: I1206 06:56:21.078806 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0af9496b-2588-4cb7-9bc2-922c48b01af2-kube-api-access\") pod \"installer-9-crc\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:21 crc kubenswrapper[4945]: I1206 06:56:21.078860 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:21 crc kubenswrapper[4945]: I1206 06:56:21.078871 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-var-lock\") pod \"installer-9-crc\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:21 crc kubenswrapper[4945]: I1206 06:56:21.078992 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:21 crc kubenswrapper[4945]: I1206 06:56:21.100608 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0af9496b-2588-4cb7-9bc2-922c48b01af2-kube-api-access\") pod \"installer-9-crc\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:21 crc kubenswrapper[4945]: I1206 06:56:21.267716 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:56:22 crc kubenswrapper[4945]: E1206 06:56:22.290767 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 06:56:22 crc kubenswrapper[4945]: E1206 06:56:22.291572 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vzjts,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-82ng8_openshift-marketplace(645fe9fd-1822-4430-8a4c-17f9eeb34e4e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 06:56:22 crc kubenswrapper[4945]: E1206 06:56:22.292847 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-82ng8" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" Dec 06 06:56:23 crc kubenswrapper[4945]: E1206 06:56:23.366005 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 06:56:23 crc kubenswrapper[4945]: E1206 06:56:23.366341 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xcmfd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-csg4k_openshift-marketplace(e315043e-3530-4af3-992b-8647e92bfa2d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 06:56:23 crc kubenswrapper[4945]: E1206 06:56:23.369016 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-csg4k" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" Dec 06 06:56:26 crc kubenswrapper[4945]: E1206 06:56:26.802617 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-csg4k" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" Dec 06 06:56:26 crc kubenswrapper[4945]: E1206 06:56:26.802791 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-82ng8" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" Dec 06 06:56:28 crc kubenswrapper[4945]: E1206 06:56:28.878250 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 06:56:28 crc kubenswrapper[4945]: E1206 06:56:28.878728 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4l5gp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-lnz2w_openshift-marketplace(a82181e2-5f33-410a-bec7-99ff14853494): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 06:56:28 crc kubenswrapper[4945]: E1206 06:56:28.879976 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-lnz2w" podUID="a82181e2-5f33-410a-bec7-99ff14853494" Dec 06 06:56:32 crc kubenswrapper[4945]: E1206 06:56:32.975361 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-lnz2w" podUID="a82181e2-5f33-410a-bec7-99ff14853494" Dec 06 06:56:34 crc kubenswrapper[4945]: E1206 06:56:34.614225 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 06:56:34 crc kubenswrapper[4945]: E1206 06:56:34.614554 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bq8mz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5w5w9_openshift-marketplace(4c50068a-69e4-4063-8f08-e4dc4e51fc8d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 06:56:34 crc kubenswrapper[4945]: E1206 06:56:34.615712 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5w5w9" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.224890 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5w5w9" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.289044 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.289205 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fgmzj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-8zwsd_openshift-marketplace(5e37b344-d572-4952-982c-8924cdf9600d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.290397 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-8zwsd" podUID="5e37b344-d572-4952-982c-8924cdf9600d" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.311916 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.312039 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8h8vw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rdjtn_openshift-marketplace(5d88a868-4630-4744-b0ae-b73f8bdf7331): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.313317 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rdjtn" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.335710 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.336028 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kd96d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xsktd_openshift-marketplace(fb6b366b-20cd-404c-be4c-b07df65de3bb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.340109 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xsktd" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.349732 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.349869 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kfw44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-fvjzs_openshift-marketplace(319506ca-cc5d-424b-a84f-e7fab30c8a76): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.351175 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-fvjzs" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" Dec 06 06:56:36 crc kubenswrapper[4945]: I1206 06:56:36.502331 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 06:56:36 crc kubenswrapper[4945]: I1206 06:56:36.667855 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 06:56:36 crc kubenswrapper[4945]: I1206 06:56:36.694128 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"99fa4df9-99a0-4cbf-9562-15bac87fa8ab","Type":"ContainerStarted","Data":"a6ba53c81dbb69539bae66766a1c3cc1c04f0604d60238966bb68256708d4905"} Dec 06 06:56:36 crc kubenswrapper[4945]: I1206 06:56:36.698366 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2xkp9" event={"ID":"d4326ced-ae97-4301-854b-222062d4764e","Type":"ContainerStarted","Data":"7c0a714ed06068029114dfcd881e63fb59dfb6e939e8e8ab8c8052c87c189f1e"} Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.703120 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-fvjzs" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.703382 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsktd" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.703971 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-8zwsd" podUID="5e37b344-d572-4952-982c-8924cdf9600d" Dec 06 06:56:36 crc kubenswrapper[4945]: E1206 06:56:36.704573 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rdjtn" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" Dec 06 06:56:36 crc kubenswrapper[4945]: I1206 06:56:36.790545 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-2xkp9" podStartSLOduration=188.790526069 podStartE2EDuration="3m8.790526069s" podCreationTimestamp="2025-12-06 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:56:36.769466163 +0000 UTC m=+210.224327207" watchObservedRunningTime="2025-12-06 06:56:36.790526069 +0000 UTC m=+210.245387113" Dec 06 06:56:37 crc kubenswrapper[4945]: I1206 06:56:37.707432 4945 generic.go:334] "Generic (PLEG): container finished" podID="99fa4df9-99a0-4cbf-9562-15bac87fa8ab" containerID="02433b9324a92d8322c0100e6fdb8d4e0971f36da75938cf7b50b65fc7b5e93f" exitCode=0 Dec 06 06:56:37 crc kubenswrapper[4945]: I1206 06:56:37.707566 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"99fa4df9-99a0-4cbf-9562-15bac87fa8ab","Type":"ContainerDied","Data":"02433b9324a92d8322c0100e6fdb8d4e0971f36da75938cf7b50b65fc7b5e93f"} Dec 06 06:56:37 crc kubenswrapper[4945]: I1206 06:56:37.711324 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0af9496b-2588-4cb7-9bc2-922c48b01af2","Type":"ContainerStarted","Data":"725ed732daa5dff45c92a819a50bfbca3f9efb4e04906fa13e207d758f7f519f"} Dec 06 06:56:37 crc kubenswrapper[4945]: I1206 06:56:37.711398 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0af9496b-2588-4cb7-9bc2-922c48b01af2","Type":"ContainerStarted","Data":"98be902e3d7e6cae8741b160756e18de2ea9898a739651b9cb6b7154b7e771db"} Dec 06 06:56:37 crc kubenswrapper[4945]: I1206 06:56:37.743431 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=17.743407211 podStartE2EDuration="17.743407211s" podCreationTimestamp="2025-12-06 06:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:56:37.741029047 +0000 UTC m=+211.195890091" watchObservedRunningTime="2025-12-06 06:56:37.743407211 +0000 UTC m=+211.198268255" Dec 06 06:56:38 crc kubenswrapper[4945]: I1206 06:56:38.796342 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:56:38 crc kubenswrapper[4945]: I1206 06:56:38.797233 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:56:38 crc kubenswrapper[4945]: I1206 06:56:38.797329 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 06:56:38 crc kubenswrapper[4945]: I1206 06:56:38.797996 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:56:38 crc kubenswrapper[4945]: I1206 06:56:38.798112 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02" gracePeriod=600 Dec 06 06:56:38 crc kubenswrapper[4945]: I1206 06:56:38.978372 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.136863 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kube-api-access\") pod \"99fa4df9-99a0-4cbf-9562-15bac87fa8ab\" (UID: \"99fa4df9-99a0-4cbf-9562-15bac87fa8ab\") " Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.137066 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kubelet-dir\") pod \"99fa4df9-99a0-4cbf-9562-15bac87fa8ab\" (UID: \"99fa4df9-99a0-4cbf-9562-15bac87fa8ab\") " Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.138669 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "99fa4df9-99a0-4cbf-9562-15bac87fa8ab" (UID: "99fa4df9-99a0-4cbf-9562-15bac87fa8ab"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.144469 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "99fa4df9-99a0-4cbf-9562-15bac87fa8ab" (UID: "99fa4df9-99a0-4cbf-9562-15bac87fa8ab"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.240097 4945 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.240516 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fa4df9-99a0-4cbf-9562-15bac87fa8ab-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.726885 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"99fa4df9-99a0-4cbf-9562-15bac87fa8ab","Type":"ContainerDied","Data":"a6ba53c81dbb69539bae66766a1c3cc1c04f0604d60238966bb68256708d4905"} Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.727712 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6ba53c81dbb69539bae66766a1c3cc1c04f0604d60238966bb68256708d4905" Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.726920 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.730250 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02" exitCode=0 Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.730423 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02"} Dec 06 06:56:39 crc kubenswrapper[4945]: I1206 06:56:39.730513 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"c0055ca88a5066f06888e8872d8e2fb3e70f24ff64dd4e6919f6a317fcc06177"} Dec 06 06:56:40 crc kubenswrapper[4945]: I1206 06:56:40.739119 4945 generic.go:334] "Generic (PLEG): container finished" podID="e315043e-3530-4af3-992b-8647e92bfa2d" containerID="1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6" exitCode=0 Dec 06 06:56:40 crc kubenswrapper[4945]: I1206 06:56:40.739222 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csg4k" event={"ID":"e315043e-3530-4af3-992b-8647e92bfa2d","Type":"ContainerDied","Data":"1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6"} Dec 06 06:56:41 crc kubenswrapper[4945]: I1206 06:56:41.747180 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csg4k" event={"ID":"e315043e-3530-4af3-992b-8647e92bfa2d","Type":"ContainerStarted","Data":"b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11"} Dec 06 06:56:41 crc kubenswrapper[4945]: I1206 06:56:41.770375 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-csg4k" podStartSLOduration=4.318619934 podStartE2EDuration="1m6.770355603s" podCreationTimestamp="2025-12-06 06:55:35 +0000 UTC" firstStartedPulling="2025-12-06 06:55:38.739558052 +0000 UTC m=+152.194419096" lastFinishedPulling="2025-12-06 06:56:41.191293721 +0000 UTC m=+214.646154765" observedRunningTime="2025-12-06 06:56:41.765516683 +0000 UTC m=+215.220377727" watchObservedRunningTime="2025-12-06 06:56:41.770355603 +0000 UTC m=+215.225216657" Dec 06 06:56:43 crc kubenswrapper[4945]: I1206 06:56:43.759687 4945 generic.go:334] "Generic (PLEG): container finished" podID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerID="aca9681621af261741913e43b4d246abf8fbf0b61eda10920cd084cd1f607873" exitCode=0 Dec 06 06:56:43 crc kubenswrapper[4945]: I1206 06:56:43.759770 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82ng8" event={"ID":"645fe9fd-1822-4430-8a4c-17f9eeb34e4e","Type":"ContainerDied","Data":"aca9681621af261741913e43b4d246abf8fbf0b61eda10920cd084cd1f607873"} Dec 06 06:56:44 crc kubenswrapper[4945]: I1206 06:56:44.781371 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82ng8" event={"ID":"645fe9fd-1822-4430-8a4c-17f9eeb34e4e","Type":"ContainerStarted","Data":"5a4f278db77f10b8fd7c56e530d8f75928edd7573b497e9eccfdf53baa158f3f"} Dec 06 06:56:44 crc kubenswrapper[4945]: I1206 06:56:44.797735 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-82ng8" podStartSLOduration=4.352460918 podStartE2EDuration="1m9.797704169s" podCreationTimestamp="2025-12-06 06:55:35 +0000 UTC" firstStartedPulling="2025-12-06 06:55:38.702517632 +0000 UTC m=+152.157378676" lastFinishedPulling="2025-12-06 06:56:44.147760883 +0000 UTC m=+217.602621927" observedRunningTime="2025-12-06 06:56:44.795956502 +0000 UTC m=+218.250817556" watchObservedRunningTime="2025-12-06 06:56:44.797704169 +0000 UTC m=+218.252565213" Dec 06 06:56:45 crc kubenswrapper[4945]: I1206 06:56:45.719711 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:56:45 crc kubenswrapper[4945]: I1206 06:56:45.719762 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:56:46 crc kubenswrapper[4945]: I1206 06:56:46.000567 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:56:46 crc kubenswrapper[4945]: I1206 06:56:46.000649 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:56:47 crc kubenswrapper[4945]: I1206 06:56:47.248894 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:56:47 crc kubenswrapper[4945]: I1206 06:56:47.249583 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:56:47 crc kubenswrapper[4945]: I1206 06:56:47.300619 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:56:48 crc kubenswrapper[4945]: I1206 06:56:48.828845 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lnz2w" event={"ID":"a82181e2-5f33-410a-bec7-99ff14853494","Type":"ContainerStarted","Data":"f929515d7650d8b6a9378d31c59e7d4b7d7b5d3d3b6e23054a18c40221421e36"} Dec 06 06:56:48 crc kubenswrapper[4945]: I1206 06:56:48.832756 4945 generic.go:334] "Generic (PLEG): container finished" podID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerID="2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37" exitCode=0 Dec 06 06:56:48 crc kubenswrapper[4945]: I1206 06:56:48.832836 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w5w9" event={"ID":"4c50068a-69e4-4063-8f08-e4dc4e51fc8d","Type":"ContainerDied","Data":"2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37"} Dec 06 06:56:49 crc kubenswrapper[4945]: I1206 06:56:49.841659 4945 generic.go:334] "Generic (PLEG): container finished" podID="a82181e2-5f33-410a-bec7-99ff14853494" containerID="f929515d7650d8b6a9378d31c59e7d4b7d7b5d3d3b6e23054a18c40221421e36" exitCode=0 Dec 06 06:56:49 crc kubenswrapper[4945]: I1206 06:56:49.841752 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lnz2w" event={"ID":"a82181e2-5f33-410a-bec7-99ff14853494","Type":"ContainerDied","Data":"f929515d7650d8b6a9378d31c59e7d4b7d7b5d3d3b6e23054a18c40221421e36"} Dec 06 06:56:49 crc kubenswrapper[4945]: I1206 06:56:49.846798 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w5w9" event={"ID":"4c50068a-69e4-4063-8f08-e4dc4e51fc8d","Type":"ContainerStarted","Data":"cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f"} Dec 06 06:56:49 crc kubenswrapper[4945]: I1206 06:56:49.849153 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvjzs" event={"ID":"319506ca-cc5d-424b-a84f-e7fab30c8a76","Type":"ContainerStarted","Data":"c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a"} Dec 06 06:56:49 crc kubenswrapper[4945]: I1206 06:56:49.884533 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5w5w9" podStartSLOduration=5.575614417 podStartE2EDuration="1m12.884511656s" podCreationTimestamp="2025-12-06 06:55:37 +0000 UTC" firstStartedPulling="2025-12-06 06:55:41.912376854 +0000 UTC m=+155.367237898" lastFinishedPulling="2025-12-06 06:56:49.221274093 +0000 UTC m=+222.676135137" observedRunningTime="2025-12-06 06:56:49.88245392 +0000 UTC m=+223.337314964" watchObservedRunningTime="2025-12-06 06:56:49.884511656 +0000 UTC m=+223.339372700" Dec 06 06:56:51 crc kubenswrapper[4945]: I1206 06:56:51.872122 4945 generic.go:334] "Generic (PLEG): container finished" podID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerID="93e12eeba540b0d0bec850c5e5034445105b5f14329fb43e7b029a5892838d4f" exitCode=0 Dec 06 06:56:51 crc kubenswrapper[4945]: I1206 06:56:51.872730 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdjtn" event={"ID":"5d88a868-4630-4744-b0ae-b73f8bdf7331","Type":"ContainerDied","Data":"93e12eeba540b0d0bec850c5e5034445105b5f14329fb43e7b029a5892838d4f"} Dec 06 06:56:51 crc kubenswrapper[4945]: I1206 06:56:51.883118 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lnz2w" event={"ID":"a82181e2-5f33-410a-bec7-99ff14853494","Type":"ContainerStarted","Data":"61398468889f56c18fbf0d95e1d5c7251628b2b76ba86ff163f4daad26ac8a5d"} Dec 06 06:56:51 crc kubenswrapper[4945]: I1206 06:56:51.887741 4945 generic.go:334] "Generic (PLEG): container finished" podID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerID="c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a" exitCode=0 Dec 06 06:56:51 crc kubenswrapper[4945]: I1206 06:56:51.887809 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvjzs" event={"ID":"319506ca-cc5d-424b-a84f-e7fab30c8a76","Type":"ContainerDied","Data":"c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a"} Dec 06 06:56:51 crc kubenswrapper[4945]: I1206 06:56:51.933996 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lnz2w" podStartSLOduration=4.865788266 podStartE2EDuration="1m13.933973319s" podCreationTimestamp="2025-12-06 06:55:38 +0000 UTC" firstStartedPulling="2025-12-06 06:55:41.88855706 +0000 UTC m=+155.343418104" lastFinishedPulling="2025-12-06 06:56:50.956742113 +0000 UTC m=+224.411603157" observedRunningTime="2025-12-06 06:56:51.932406317 +0000 UTC m=+225.387267371" watchObservedRunningTime="2025-12-06 06:56:51.933973319 +0000 UTC m=+225.388834363" Dec 06 06:56:52 crc kubenswrapper[4945]: I1206 06:56:52.898205 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsktd" event={"ID":"fb6b366b-20cd-404c-be4c-b07df65de3bb","Type":"ContainerStarted","Data":"ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf"} Dec 06 06:56:52 crc kubenswrapper[4945]: I1206 06:56:52.901164 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvjzs" event={"ID":"319506ca-cc5d-424b-a84f-e7fab30c8a76","Type":"ContainerStarted","Data":"a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0"} Dec 06 06:56:52 crc kubenswrapper[4945]: I1206 06:56:52.903940 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zwsd" event={"ID":"5e37b344-d572-4952-982c-8924cdf9600d","Type":"ContainerStarted","Data":"53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10"} Dec 06 06:56:52 crc kubenswrapper[4945]: I1206 06:56:52.908195 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdjtn" event={"ID":"5d88a868-4630-4744-b0ae-b73f8bdf7331","Type":"ContainerStarted","Data":"024c5e38f678aea6f15e8f0dde651f77692d1f1a769ece0654193b0adccc6745"} Dec 06 06:56:52 crc kubenswrapper[4945]: I1206 06:56:52.988809 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rdjtn" podStartSLOduration=4.620139147 podStartE2EDuration="1m14.988783737s" podCreationTimestamp="2025-12-06 06:55:38 +0000 UTC" firstStartedPulling="2025-12-06 06:55:41.93360946 +0000 UTC m=+155.388470504" lastFinishedPulling="2025-12-06 06:56:52.30225405 +0000 UTC m=+225.757115094" observedRunningTime="2025-12-06 06:56:52.964938667 +0000 UTC m=+226.419799721" watchObservedRunningTime="2025-12-06 06:56:52.988783737 +0000 UTC m=+226.443644791" Dec 06 06:56:53 crc kubenswrapper[4945]: I1206 06:56:53.016030 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fvjzs" podStartSLOduration=4.579511237 podStartE2EDuration="1m15.016012099s" podCreationTimestamp="2025-12-06 06:55:38 +0000 UTC" firstStartedPulling="2025-12-06 06:55:41.917547449 +0000 UTC m=+155.372408493" lastFinishedPulling="2025-12-06 06:56:52.354048311 +0000 UTC m=+225.808909355" observedRunningTime="2025-12-06 06:56:53.014110037 +0000 UTC m=+226.468971091" watchObservedRunningTime="2025-12-06 06:56:53.016012099 +0000 UTC m=+226.470873143" Dec 06 06:56:53 crc kubenswrapper[4945]: I1206 06:56:53.916797 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e37b344-d572-4952-982c-8924cdf9600d" containerID="53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10" exitCode=0 Dec 06 06:56:53 crc kubenswrapper[4945]: I1206 06:56:53.917247 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zwsd" event={"ID":"5e37b344-d572-4952-982c-8924cdf9600d","Type":"ContainerDied","Data":"53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10"} Dec 06 06:56:53 crc kubenswrapper[4945]: I1206 06:56:53.921421 4945 generic.go:334] "Generic (PLEG): container finished" podID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerID="ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf" exitCode=0 Dec 06 06:56:53 crc kubenswrapper[4945]: I1206 06:56:53.921477 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsktd" event={"ID":"fb6b366b-20cd-404c-be4c-b07df65de3bb","Type":"ContainerDied","Data":"ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf"} Dec 06 06:56:56 crc kubenswrapper[4945]: I1206 06:56:56.040100 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:56:57 crc kubenswrapper[4945]: I1206 06:56:57.106090 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-82ng8"] Dec 06 06:56:57 crc kubenswrapper[4945]: I1206 06:56:57.106349 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-82ng8" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerName="registry-server" containerID="cri-o://5a4f278db77f10b8fd7c56e530d8f75928edd7573b497e9eccfdf53baa158f3f" gracePeriod=2 Dec 06 06:56:57 crc kubenswrapper[4945]: I1206 06:56:57.588034 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:56:57 crc kubenswrapper[4945]: I1206 06:56:57.588495 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:56:57 crc kubenswrapper[4945]: I1206 06:56:57.633589 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:56:57 crc kubenswrapper[4945]: I1206 06:56:57.967328 4945 generic.go:334] "Generic (PLEG): container finished" podID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerID="5a4f278db77f10b8fd7c56e530d8f75928edd7573b497e9eccfdf53baa158f3f" exitCode=0 Dec 06 06:56:57 crc kubenswrapper[4945]: I1206 06:56:57.967508 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82ng8" event={"ID":"645fe9fd-1822-4430-8a4c-17f9eeb34e4e","Type":"ContainerDied","Data":"5a4f278db77f10b8fd7c56e530d8f75928edd7573b497e9eccfdf53baa158f3f"} Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.015363 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.161260 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.169895 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzjts\" (UniqueName: \"kubernetes.io/projected/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-kube-api-access-vzjts\") pod \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.169948 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-utilities\") pod \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.169986 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-catalog-content\") pod \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\" (UID: \"645fe9fd-1822-4430-8a4c-17f9eeb34e4e\") " Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.171174 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-utilities" (OuterVolumeSpecName: "utilities") pod "645fe9fd-1822-4430-8a4c-17f9eeb34e4e" (UID: "645fe9fd-1822-4430-8a4c-17f9eeb34e4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.180586 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-kube-api-access-vzjts" (OuterVolumeSpecName: "kube-api-access-vzjts") pod "645fe9fd-1822-4430-8a4c-17f9eeb34e4e" (UID: "645fe9fd-1822-4430-8a4c-17f9eeb34e4e"). InnerVolumeSpecName "kube-api-access-vzjts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.254707 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "645fe9fd-1822-4430-8a4c-17f9eeb34e4e" (UID: "645fe9fd-1822-4430-8a4c-17f9eeb34e4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.270955 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzjts\" (UniqueName: \"kubernetes.io/projected/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-kube-api-access-vzjts\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.270990 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.271002 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/645fe9fd-1822-4430-8a4c-17f9eeb34e4e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.815619 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.815992 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.839273 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.839349 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.858866 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.886207 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.975850 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zwsd" event={"ID":"5e37b344-d572-4952-982c-8924cdf9600d","Type":"ContainerStarted","Data":"6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb"} Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.978816 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82ng8" event={"ID":"645fe9fd-1822-4430-8a4c-17f9eeb34e4e","Type":"ContainerDied","Data":"22ddf88a4070c220690bf412892c856d3c5b8ad7289c53d9ec181bd59823cd0f"} Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.979342 4945 scope.go:117] "RemoveContainer" containerID="5a4f278db77f10b8fd7c56e530d8f75928edd7573b497e9eccfdf53baa158f3f" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.978911 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82ng8" Dec 06 06:56:58 crc kubenswrapper[4945]: I1206 06:56:58.981544 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsktd" event={"ID":"fb6b366b-20cd-404c-be4c-b07df65de3bb","Type":"ContainerStarted","Data":"414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64"} Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.001431 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8zwsd" podStartSLOduration=5.8468983439999995 podStartE2EDuration="1m24.0014132s" podCreationTimestamp="2025-12-06 06:55:35 +0000 UTC" firstStartedPulling="2025-12-06 06:55:39.850648907 +0000 UTC m=+153.305509951" lastFinishedPulling="2025-12-06 06:56:58.005163763 +0000 UTC m=+231.460024807" observedRunningTime="2025-12-06 06:56:59.000365962 +0000 UTC m=+232.455226996" watchObservedRunningTime="2025-12-06 06:56:59.0014132 +0000 UTC m=+232.456274244" Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.005457 4945 scope.go:117] "RemoveContainer" containerID="aca9681621af261741913e43b4d246abf8fbf0b61eda10920cd084cd1f607873" Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.019553 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-82ng8"] Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.022579 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-82ng8"] Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.028849 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.033023 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.034616 4945 scope.go:117] "RemoveContainer" containerID="7f8f8c1b1c00f5b742aa19e066979f4fb9c509d9ebbbe7afbb70a4285acda9bb" Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.064464 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xsktd" podStartSLOduration=4.890514734 podStartE2EDuration="1m24.064448113s" podCreationTimestamp="2025-12-06 06:55:35 +0000 UTC" firstStartedPulling="2025-12-06 06:55:38.724244681 +0000 UTC m=+152.179105725" lastFinishedPulling="2025-12-06 06:56:57.89817806 +0000 UTC m=+231.353039104" observedRunningTime="2025-12-06 06:56:59.039703168 +0000 UTC m=+232.494564212" watchObservedRunningTime="2025-12-06 06:56:59.064448113 +0000 UTC m=+232.519309157" Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.192629 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.192698 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:56:59 crc kubenswrapper[4945]: I1206 06:56:59.236836 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:57:00 crc kubenswrapper[4945]: I1206 06:57:00.033221 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:57:00 crc kubenswrapper[4945]: I1206 06:57:00.964007 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" path="/var/lib/kubelet/pods/645fe9fd-1822-4430-8a4c-17f9eeb34e4e/volumes" Dec 06 06:57:01 crc kubenswrapper[4945]: I1206 06:57:01.509103 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdjtn"] Dec 06 06:57:01 crc kubenswrapper[4945]: I1206 06:57:01.509453 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rdjtn" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerName="registry-server" containerID="cri-o://024c5e38f678aea6f15e8f0dde651f77692d1f1a769ece0654193b0adccc6745" gracePeriod=2 Dec 06 06:57:02 crc kubenswrapper[4945]: I1206 06:57:02.919734 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lnz2w"] Dec 06 06:57:02 crc kubenswrapper[4945]: I1206 06:57:02.920626 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lnz2w" podUID="a82181e2-5f33-410a-bec7-99ff14853494" containerName="registry-server" containerID="cri-o://61398468889f56c18fbf0d95e1d5c7251628b2b76ba86ff163f4daad26ac8a5d" gracePeriod=2 Dec 06 06:57:05 crc kubenswrapper[4945]: I1206 06:57:05.768661 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:57:05 crc kubenswrapper[4945]: I1206 06:57:05.769185 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:57:05 crc kubenswrapper[4945]: I1206 06:57:05.812855 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.033307 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rdjtn_5d88a868-4630-4744-b0ae-b73f8bdf7331/registry-server/0.log" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.034689 4945 generic.go:334] "Generic (PLEG): container finished" podID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerID="024c5e38f678aea6f15e8f0dde651f77692d1f1a769ece0654193b0adccc6745" exitCode=137 Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.034823 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdjtn" event={"ID":"5d88a868-4630-4744-b0ae-b73f8bdf7331","Type":"ContainerDied","Data":"024c5e38f678aea6f15e8f0dde651f77692d1f1a769ece0654193b0adccc6745"} Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.038447 4945 generic.go:334] "Generic (PLEG): container finished" podID="a82181e2-5f33-410a-bec7-99ff14853494" containerID="61398468889f56c18fbf0d95e1d5c7251628b2b76ba86ff163f4daad26ac8a5d" exitCode=0 Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.038493 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lnz2w" event={"ID":"a82181e2-5f33-410a-bec7-99ff14853494","Type":"ContainerDied","Data":"61398468889f56c18fbf0d95e1d5c7251628b2b76ba86ff163f4daad26ac8a5d"} Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.093345 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.254089 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.254181 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.324770 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.697780 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rdjtn_5d88a868-4630-4744-b0ae-b73f8bdf7331/registry-server/0.log" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.699799 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.703500 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.817073 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-utilities\") pod \"a82181e2-5f33-410a-bec7-99ff14853494\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.817185 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-catalog-content\") pod \"a82181e2-5f33-410a-bec7-99ff14853494\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.817302 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-utilities\") pod \"5d88a868-4630-4744-b0ae-b73f8bdf7331\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.817378 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l5gp\" (UniqueName: \"kubernetes.io/projected/a82181e2-5f33-410a-bec7-99ff14853494-kube-api-access-4l5gp\") pod \"a82181e2-5f33-410a-bec7-99ff14853494\" (UID: \"a82181e2-5f33-410a-bec7-99ff14853494\") " Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.817417 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-catalog-content\") pod \"5d88a868-4630-4744-b0ae-b73f8bdf7331\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.817442 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h8vw\" (UniqueName: \"kubernetes.io/projected/5d88a868-4630-4744-b0ae-b73f8bdf7331-kube-api-access-8h8vw\") pod \"5d88a868-4630-4744-b0ae-b73f8bdf7331\" (UID: \"5d88a868-4630-4744-b0ae-b73f8bdf7331\") " Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.818140 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-utilities" (OuterVolumeSpecName: "utilities") pod "5d88a868-4630-4744-b0ae-b73f8bdf7331" (UID: "5d88a868-4630-4744-b0ae-b73f8bdf7331"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.818804 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-utilities" (OuterVolumeSpecName: "utilities") pod "a82181e2-5f33-410a-bec7-99ff14853494" (UID: "a82181e2-5f33-410a-bec7-99ff14853494"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.825507 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a82181e2-5f33-410a-bec7-99ff14853494-kube-api-access-4l5gp" (OuterVolumeSpecName: "kube-api-access-4l5gp") pod "a82181e2-5f33-410a-bec7-99ff14853494" (UID: "a82181e2-5f33-410a-bec7-99ff14853494"). InnerVolumeSpecName "kube-api-access-4l5gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.826071 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d88a868-4630-4744-b0ae-b73f8bdf7331-kube-api-access-8h8vw" (OuterVolumeSpecName: "kube-api-access-8h8vw") pod "5d88a868-4630-4744-b0ae-b73f8bdf7331" (UID: "5d88a868-4630-4744-b0ae-b73f8bdf7331"). InnerVolumeSpecName "kube-api-access-8h8vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.837157 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d88a868-4630-4744-b0ae-b73f8bdf7331" (UID: "5d88a868-4630-4744-b0ae-b73f8bdf7331"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.918672 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l5gp\" (UniqueName: \"kubernetes.io/projected/a82181e2-5f33-410a-bec7-99ff14853494-kube-api-access-4l5gp\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.918712 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.918723 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h8vw\" (UniqueName: \"kubernetes.io/projected/5d88a868-4630-4744-b0ae-b73f8bdf7331-kube-api-access-8h8vw\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.918732 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.918740 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d88a868-4630-4744-b0ae-b73f8bdf7331-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:06 crc kubenswrapper[4945]: I1206 06:57:06.933506 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a82181e2-5f33-410a-bec7-99ff14853494" (UID: "a82181e2-5f33-410a-bec7-99ff14853494"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.019693 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a82181e2-5f33-410a-bec7-99ff14853494-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.047916 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdjtn" event={"ID":"5d88a868-4630-4744-b0ae-b73f8bdf7331","Type":"ContainerDied","Data":"605db9afa1aa66d7345821180794d0a566bafc1e5a1edcd919d02a33ee0651e8"} Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.047983 4945 scope.go:117] "RemoveContainer" containerID="024c5e38f678aea6f15e8f0dde651f77692d1f1a769ece0654193b0adccc6745" Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.047998 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdjtn" Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.052794 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lnz2w" event={"ID":"a82181e2-5f33-410a-bec7-99ff14853494","Type":"ContainerDied","Data":"e3b27405e990eef62baa555435c6531ae0b2e71158e9093b9c055a63c0c116c2"} Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.052905 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lnz2w" Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.074018 4945 scope.go:117] "RemoveContainer" containerID="93e12eeba540b0d0bec850c5e5034445105b5f14329fb43e7b029a5892838d4f" Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.096676 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lnz2w"] Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.100775 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lnz2w"] Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.109809 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdjtn"] Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.112180 4945 scope.go:117] "RemoveContainer" containerID="38295fbf9d19f23c736d583ec5a9bd6d5350c0f352da8bcc6c70e8db76fedacd" Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.115487 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdjtn"] Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.119922 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.133924 4945 scope.go:117] "RemoveContainer" containerID="61398468889f56c18fbf0d95e1d5c7251628b2b76ba86ff163f4daad26ac8a5d" Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.165128 4945 scope.go:117] "RemoveContainer" containerID="f929515d7650d8b6a9378d31c59e7d4b7d7b5d3d3b6e23054a18c40221421e36" Dec 06 06:57:07 crc kubenswrapper[4945]: I1206 06:57:07.184548 4945 scope.go:117] "RemoveContainer" containerID="cfa6c17f0d891f3171a04a79c8106f794990b88e381a516d227c1a0da4060ba8" Dec 06 06:57:08 crc kubenswrapper[4945]: I1206 06:57:08.753586 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-456vw"] Dec 06 06:57:08 crc kubenswrapper[4945]: I1206 06:57:08.962046 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" path="/var/lib/kubelet/pods/5d88a868-4630-4744-b0ae-b73f8bdf7331/volumes" Dec 06 06:57:08 crc kubenswrapper[4945]: I1206 06:57:08.962958 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a82181e2-5f33-410a-bec7-99ff14853494" path="/var/lib/kubelet/pods/a82181e2-5f33-410a-bec7-99ff14853494/volumes" Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.306659 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8zwsd"] Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.307129 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8zwsd" podUID="5e37b344-d572-4952-982c-8924cdf9600d" containerName="registry-server" containerID="cri-o://6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb" gracePeriod=2 Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.698817 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.859519 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgmzj\" (UniqueName: \"kubernetes.io/projected/5e37b344-d572-4952-982c-8924cdf9600d-kube-api-access-fgmzj\") pod \"5e37b344-d572-4952-982c-8924cdf9600d\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.860437 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-catalog-content\") pod \"5e37b344-d572-4952-982c-8924cdf9600d\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.860596 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-utilities\") pod \"5e37b344-d572-4952-982c-8924cdf9600d\" (UID: \"5e37b344-d572-4952-982c-8924cdf9600d\") " Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.862270 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-utilities" (OuterVolumeSpecName: "utilities") pod "5e37b344-d572-4952-982c-8924cdf9600d" (UID: "5e37b344-d572-4952-982c-8924cdf9600d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.869380 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e37b344-d572-4952-982c-8924cdf9600d-kube-api-access-fgmzj" (OuterVolumeSpecName: "kube-api-access-fgmzj") pod "5e37b344-d572-4952-982c-8924cdf9600d" (UID: "5e37b344-d572-4952-982c-8924cdf9600d"). InnerVolumeSpecName "kube-api-access-fgmzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.914114 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e37b344-d572-4952-982c-8924cdf9600d" (UID: "5e37b344-d572-4952-982c-8924cdf9600d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.962719 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgmzj\" (UniqueName: \"kubernetes.io/projected/5e37b344-d572-4952-982c-8924cdf9600d-kube-api-access-fgmzj\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.962762 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:09 crc kubenswrapper[4945]: I1206 06:57:09.962774 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e37b344-d572-4952-982c-8924cdf9600d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.088670 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e37b344-d572-4952-982c-8924cdf9600d" containerID="6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb" exitCode=0 Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.088766 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zwsd" event={"ID":"5e37b344-d572-4952-982c-8924cdf9600d","Type":"ContainerDied","Data":"6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb"} Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.088786 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zwsd" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.088832 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zwsd" event={"ID":"5e37b344-d572-4952-982c-8924cdf9600d","Type":"ContainerDied","Data":"f62a1839e5b093efc9c68f6b10260e6572723f04e1d4e32ec1713c4fc70fccc9"} Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.088873 4945 scope.go:117] "RemoveContainer" containerID="6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.108999 4945 scope.go:117] "RemoveContainer" containerID="53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.126854 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8zwsd"] Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.129823 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8zwsd"] Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.140778 4945 scope.go:117] "RemoveContainer" containerID="a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.160848 4945 scope.go:117] "RemoveContainer" containerID="6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb" Dec 06 06:57:10 crc kubenswrapper[4945]: E1206 06:57:10.161330 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb\": container with ID starting with 6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb not found: ID does not exist" containerID="6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.161368 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb"} err="failed to get container status \"6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb\": rpc error: code = NotFound desc = could not find container \"6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb\": container with ID starting with 6a4e1ebc659b263e279ee0a5a480d543e46728e88e0d0c740318bfca15e610fb not found: ID does not exist" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.161403 4945 scope.go:117] "RemoveContainer" containerID="53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10" Dec 06 06:57:10 crc kubenswrapper[4945]: E1206 06:57:10.161812 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10\": container with ID starting with 53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10 not found: ID does not exist" containerID="53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.161835 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10"} err="failed to get container status \"53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10\": rpc error: code = NotFound desc = could not find container \"53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10\": container with ID starting with 53ee3ab109ca8ed1e1aaf05395ef01b2226e43784acea4734e1e55d5ae3e9e10 not found: ID does not exist" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.161851 4945 scope.go:117] "RemoveContainer" containerID="a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198" Dec 06 06:57:10 crc kubenswrapper[4945]: E1206 06:57:10.162086 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198\": container with ID starting with a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198 not found: ID does not exist" containerID="a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.162116 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198"} err="failed to get container status \"a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198\": rpc error: code = NotFound desc = could not find container \"a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198\": container with ID starting with a888d16ad6cb838a70a78bed55633f8095fd47fe23eb8220ec15be4eb8605198 not found: ID does not exist" Dec 06 06:57:10 crc kubenswrapper[4945]: I1206 06:57:10.964050 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e37b344-d572-4952-982c-8924cdf9600d" path="/var/lib/kubelet/pods/5e37b344-d572-4952-982c-8924cdf9600d/volumes" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.756006 4945 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.757547 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5" gracePeriod=15 Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.757582 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5" gracePeriod=15 Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.757748 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb" gracePeriod=15 Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.757799 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2" gracePeriod=15 Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.757841 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41" gracePeriod=15 Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.758912 4945 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759380 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82181e2-5f33-410a-bec7-99ff14853494" containerName="extract-utilities" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759412 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82181e2-5f33-410a-bec7-99ff14853494" containerName="extract-utilities" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759433 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerName="extract-content" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759450 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerName="extract-content" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759472 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759487 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759505 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759523 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759548 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759579 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759606 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e37b344-d572-4952-982c-8924cdf9600d" containerName="extract-content" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759622 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e37b344-d572-4952-982c-8924cdf9600d" containerName="extract-content" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759642 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e37b344-d572-4952-982c-8924cdf9600d" containerName="extract-utilities" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759657 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e37b344-d572-4952-982c-8924cdf9600d" containerName="extract-utilities" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759674 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759690 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759713 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerName="extract-utilities" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759729 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerName="extract-utilities" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759757 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82181e2-5f33-410a-bec7-99ff14853494" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759775 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82181e2-5f33-410a-bec7-99ff14853494" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759808 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759823 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.759846 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerName="extract-utilities" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.759863 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerName="extract-utilities" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.760254 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760320 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.760345 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82181e2-5f33-410a-bec7-99ff14853494" containerName="extract-content" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760362 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82181e2-5f33-410a-bec7-99ff14853494" containerName="extract-content" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.760410 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760426 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.760447 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerName="extract-content" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760463 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerName="extract-content" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.760482 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99fa4df9-99a0-4cbf-9562-15bac87fa8ab" containerName="pruner" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760500 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="99fa4df9-99a0-4cbf-9562-15bac87fa8ab" containerName="pruner" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.760524 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e37b344-d572-4952-982c-8924cdf9600d" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760540 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e37b344-d572-4952-982c-8924cdf9600d" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.760561 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760577 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760831 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d88a868-4630-4744-b0ae-b73f8bdf7331" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760861 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="645fe9fd-1822-4430-8a4c-17f9eeb34e4e" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760881 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760910 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760929 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760952 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760970 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="99fa4df9-99a0-4cbf-9562-15bac87fa8ab" containerName="pruner" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.760992 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.761012 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a82181e2-5f33-410a-bec7-99ff14853494" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.761027 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e37b344-d572-4952-982c-8924cdf9600d" containerName="registry-server" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.761050 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.761677 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.761726 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.765588 4945 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.768910 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.788246 4945 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 06 06:57:14 crc kubenswrapper[4945]: E1206 06:57:14.836088 4945 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.220:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.936833 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.936907 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.936932 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.937085 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.937161 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.937204 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.937266 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:14 crc kubenswrapper[4945]: I1206 06:57:14.937386 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039012 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039172 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039199 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039231 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039266 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039320 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039330 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039353 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039368 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039364 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039380 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039404 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039426 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039449 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039528 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.039744 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.126512 4945 generic.go:334] "Generic (PLEG): container finished" podID="0af9496b-2588-4cb7-9bc2-922c48b01af2" containerID="725ed732daa5dff45c92a819a50bfbca3f9efb4e04906fa13e207d758f7f519f" exitCode=0 Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.126682 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0af9496b-2588-4cb7-9bc2-922c48b01af2","Type":"ContainerDied","Data":"725ed732daa5dff45c92a819a50bfbca3f9efb4e04906fa13e207d758f7f519f"} Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.128064 4945 status_manager.go:851] "Failed to get status for pod" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.129929 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.131817 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.133007 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5" exitCode=0 Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.133036 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb" exitCode=0 Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.133048 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2" exitCode=0 Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.133060 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41" exitCode=2 Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.133109 4945 scope.go:117] "RemoveContainer" containerID="67ef8d8fff7acec18c756c55cffa1d0e305383fdfce5d292ef97649ace8cf770" Dec 06 06:57:15 crc kubenswrapper[4945]: I1206 06:57:15.138617 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:15 crc kubenswrapper[4945]: E1206 06:57:15.179269 4945 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.220:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e8e0020d9235f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 06:57:15.178058591 +0000 UTC m=+248.632919675,LastTimestamp:2025-12-06 06:57:15.178058591 +0000 UTC m=+248.632919675,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 06:57:15 crc kubenswrapper[4945]: E1206 06:57:15.631187 4945 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.220:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e8e0020d9235f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 06:57:15.178058591 +0000 UTC m=+248.632919675,LastTimestamp:2025-12-06 06:57:15.178058591 +0000 UTC m=+248.632919675,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.150855 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b152fc0d3dfcf1aa5cd5c20554a27524505731efeef856c8069e75a3381eaafe"} Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.150959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"086c19ce399a812db9326c12c617313ec4b4405552ba8cf70ad68406d719985f"} Dec 06 06:57:16 crc kubenswrapper[4945]: E1206 06:57:16.152262 4945 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.220:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.152384 4945 status_manager.go:851] "Failed to get status for pod" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.157843 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.449066 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.450256 4945 status_manager.go:851] "Failed to get status for pod" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.562743 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0af9496b-2588-4cb7-9bc2-922c48b01af2-kube-api-access\") pod \"0af9496b-2588-4cb7-9bc2-922c48b01af2\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.562839 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-var-lock\") pod \"0af9496b-2588-4cb7-9bc2-922c48b01af2\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.563013 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-kubelet-dir\") pod \"0af9496b-2588-4cb7-9bc2-922c48b01af2\" (UID: \"0af9496b-2588-4cb7-9bc2-922c48b01af2\") " Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.563451 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-var-lock" (OuterVolumeSpecName: "var-lock") pod "0af9496b-2588-4cb7-9bc2-922c48b01af2" (UID: "0af9496b-2588-4cb7-9bc2-922c48b01af2"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.563869 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0af9496b-2588-4cb7-9bc2-922c48b01af2" (UID: "0af9496b-2588-4cb7-9bc2-922c48b01af2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.571679 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0af9496b-2588-4cb7-9bc2-922c48b01af2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0af9496b-2588-4cb7-9bc2-922c48b01af2" (UID: "0af9496b-2588-4cb7-9bc2-922c48b01af2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.664945 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0af9496b-2588-4cb7-9bc2-922c48b01af2-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.665022 4945 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.665034 4945 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0af9496b-2588-4cb7-9bc2-922c48b01af2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:16 crc kubenswrapper[4945]: I1206 06:57:16.955767 4945 status_manager.go:851] "Failed to get status for pod" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:16 crc kubenswrapper[4945]: E1206 06:57:16.969292 4945 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.129.56.220:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" volumeName="registry-storage" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.168818 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.168815 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0af9496b-2588-4cb7-9bc2-922c48b01af2","Type":"ContainerDied","Data":"98be902e3d7e6cae8741b160756e18de2ea9898a739651b9cb6b7154b7e771db"} Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.168963 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98be902e3d7e6cae8741b160756e18de2ea9898a739651b9cb6b7154b7e771db" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.175880 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.177330 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5" exitCode=0 Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.177405 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccaf076f5bedc56719a5be66989bbc17107fe8d3952df16c5de86cc870ac3b7b" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.179829 4945 status_manager.go:851] "Failed to get status for pod" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.182208 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.183783 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.184581 4945 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.184949 4945 status_manager.go:851] "Failed to get status for pod" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.373762 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.373993 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.374033 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.374425 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.374458 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.374490 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:57:17 crc kubenswrapper[4945]: E1206 06:57:17.381013 4945 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:17 crc kubenswrapper[4945]: E1206 06:57:17.382015 4945 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:17 crc kubenswrapper[4945]: E1206 06:57:17.382911 4945 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:17 crc kubenswrapper[4945]: E1206 06:57:17.383479 4945 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:17 crc kubenswrapper[4945]: E1206 06:57:17.383887 4945 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.383933 4945 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 06 06:57:17 crc kubenswrapper[4945]: E1206 06:57:17.384374 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="200ms" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.475940 4945 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.475989 4945 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:17 crc kubenswrapper[4945]: I1206 06:57:17.476003 4945 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:17 crc kubenswrapper[4945]: E1206 06:57:17.585813 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="400ms" Dec 06 06:57:17 crc kubenswrapper[4945]: E1206 06:57:17.986895 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="800ms" Dec 06 06:57:18 crc kubenswrapper[4945]: I1206 06:57:18.183231 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:18 crc kubenswrapper[4945]: I1206 06:57:18.202051 4945 status_manager.go:851] "Failed to get status for pod" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:18 crc kubenswrapper[4945]: I1206 06:57:18.203557 4945 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:18 crc kubenswrapper[4945]: E1206 06:57:18.787683 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="1.6s" Dec 06 06:57:18 crc kubenswrapper[4945]: I1206 06:57:18.961350 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 06 06:57:20 crc kubenswrapper[4945]: E1206 06:57:20.389404 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="3.2s" Dec 06 06:57:23 crc kubenswrapper[4945]: E1206 06:57:23.590590 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.220:6443: connect: connection refused" interval="6.4s" Dec 06 06:57:25 crc kubenswrapper[4945]: E1206 06:57:25.632694 4945 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.220:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e8e0020d9235f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 06:57:15.178058591 +0000 UTC m=+248.632919675,LastTimestamp:2025-12-06 06:57:15.178058591 +0000 UTC m=+248.632919675,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 06:57:26 crc kubenswrapper[4945]: I1206 06:57:26.952604 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:26 crc kubenswrapper[4945]: I1206 06:57:26.956245 4945 status_manager.go:851] "Failed to get status for pod" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:26 crc kubenswrapper[4945]: I1206 06:57:26.957180 4945 status_manager.go:851] "Failed to get status for pod" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:26 crc kubenswrapper[4945]: I1206 06:57:26.971621 4945 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2b924ff1-dd26-4e47-b90c-144eae88b737" Dec 06 06:57:26 crc kubenswrapper[4945]: I1206 06:57:26.971666 4945 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2b924ff1-dd26-4e47-b90c-144eae88b737" Dec 06 06:57:26 crc kubenswrapper[4945]: E1206 06:57:26.972378 4945 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:26 crc kubenswrapper[4945]: I1206 06:57:26.973067 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:27 crc kubenswrapper[4945]: I1206 06:57:27.239452 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"45eee56c0ef725b90630837e22779c1ac1877defca86eb9a550403621559e24d"} Dec 06 06:57:28 crc kubenswrapper[4945]: I1206 06:57:28.246629 4945 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="3a3d63b5e6f43aae549d9c2a3b4edb0b3178e8317ef651ccd355c58852cb5129" exitCode=0 Dec 06 06:57:28 crc kubenswrapper[4945]: I1206 06:57:28.246675 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"3a3d63b5e6f43aae549d9c2a3b4edb0b3178e8317ef651ccd355c58852cb5129"} Dec 06 06:57:28 crc kubenswrapper[4945]: I1206 06:57:28.246926 4945 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2b924ff1-dd26-4e47-b90c-144eae88b737" Dec 06 06:57:28 crc kubenswrapper[4945]: I1206 06:57:28.246957 4945 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2b924ff1-dd26-4e47-b90c-144eae88b737" Dec 06 06:57:28 crc kubenswrapper[4945]: I1206 06:57:28.247222 4945 status_manager.go:851] "Failed to get status for pod" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" Dec 06 06:57:28 crc kubenswrapper[4945]: E1206 06:57:28.247424 4945 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.220:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:29 crc kubenswrapper[4945]: I1206 06:57:29.256514 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ff671e25341caf9cfe7fcb2e01920113e26e7939aa28ac6f11878352e4e037f5"} Dec 06 06:57:29 crc kubenswrapper[4945]: I1206 06:57:29.256869 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c02cf1e726a44553e08ecfa7db0fded42aa207448d07227c92403131fd8c56cb"} Dec 06 06:57:29 crc kubenswrapper[4945]: I1206 06:57:29.256887 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8b4e818f8be8bc44d5afb41bebe3cd4f31e6f10e84e85d5005586ea59967d8d1"} Dec 06 06:57:29 crc kubenswrapper[4945]: I1206 06:57:29.256898 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f13ff43596e5257cea742b28b81900bcf25c7af360922155dbdd59407c6cdb36"} Dec 06 06:57:30 crc kubenswrapper[4945]: I1206 06:57:30.265083 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 06:57:30 crc kubenswrapper[4945]: I1206 06:57:30.265416 4945 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4" exitCode=1 Dec 06 06:57:30 crc kubenswrapper[4945]: I1206 06:57:30.265501 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4"} Dec 06 06:57:30 crc kubenswrapper[4945]: I1206 06:57:30.266022 4945 scope.go:117] "RemoveContainer" containerID="30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4" Dec 06 06:57:30 crc kubenswrapper[4945]: I1206 06:57:30.269610 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"35edd5dbab2e75b975ee219973a3418134464384b1fe740a675d44944abb2c13"} Dec 06 06:57:30 crc kubenswrapper[4945]: I1206 06:57:30.270005 4945 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2b924ff1-dd26-4e47-b90c-144eae88b737" Dec 06 06:57:30 crc kubenswrapper[4945]: I1206 06:57:30.270051 4945 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2b924ff1-dd26-4e47-b90c-144eae88b737" Dec 06 06:57:30 crc kubenswrapper[4945]: I1206 06:57:30.270052 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:31 crc kubenswrapper[4945]: I1206 06:57:31.280255 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 06:57:31 crc kubenswrapper[4945]: I1206 06:57:31.280619 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"99d1d295024de88c141b9f8d7b99f57b0ce324abf4c7370dd527519050c29df4"} Dec 06 06:57:31 crc kubenswrapper[4945]: I1206 06:57:31.974085 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:31 crc kubenswrapper[4945]: I1206 06:57:31.974141 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:31 crc kubenswrapper[4945]: I1206 06:57:31.980041 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:33 crc kubenswrapper[4945]: I1206 06:57:33.639519 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:57:33 crc kubenswrapper[4945]: I1206 06:57:33.639869 4945 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 06 06:57:33 crc kubenswrapper[4945]: I1206 06:57:33.640016 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 06 06:57:33 crc kubenswrapper[4945]: I1206 06:57:33.807943 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" podUID="957f12af-ad21-4c4a-97a2-ab1b824aa5a0" containerName="oauth-openshift" containerID="cri-o://58855a5cb5145e6544f569cb20d94290c3da7d7fdae7b3429bfec3d21f8c4a95" gracePeriod=15 Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.304205 4945 generic.go:334] "Generic (PLEG): container finished" podID="957f12af-ad21-4c4a-97a2-ab1b824aa5a0" containerID="58855a5cb5145e6544f569cb20d94290c3da7d7fdae7b3429bfec3d21f8c4a95" exitCode=0 Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.304311 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" event={"ID":"957f12af-ad21-4c4a-97a2-ab1b824aa5a0","Type":"ContainerDied","Data":"58855a5cb5145e6544f569cb20d94290c3da7d7fdae7b3429bfec3d21f8c4a95"} Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.466070 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510178 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-session\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510251 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-service-ca\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510300 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-login\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510363 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzpw6\" (UniqueName: \"kubernetes.io/projected/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-kube-api-access-jzpw6\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510411 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-ocp-branding-template\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510437 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-router-certs\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510462 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-dir\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510517 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-policies\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510574 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-serving-cert\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510605 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-trusted-ca-bundle\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510617 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510643 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-idp-0-file-data\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510740 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-cliconfig\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510794 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-error\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.510832 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-provider-selection\") pod \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\" (UID: \"957f12af-ad21-4c4a-97a2-ab1b824aa5a0\") " Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.511291 4945 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.511797 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.513387 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.513376 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.514100 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.519544 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.519665 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-kube-api-access-jzpw6" (OuterVolumeSpecName: "kube-api-access-jzpw6") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "kube-api-access-jzpw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.520214 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.520772 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.520990 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.521181 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.521444 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.521703 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.522031 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "957f12af-ad21-4c4a-97a2-ab1b824aa5a0" (UID: "957f12af-ad21-4c4a-97a2-ab1b824aa5a0"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612717 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612757 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612770 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612782 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzpw6\" (UniqueName: \"kubernetes.io/projected/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-kube-api-access-jzpw6\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612796 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612805 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612815 4945 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612825 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612834 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612843 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612852 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612860 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:34 crc kubenswrapper[4945]: I1206 06:57:34.612870 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/957f12af-ad21-4c4a-97a2-ab1b824aa5a0-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:35 crc kubenswrapper[4945]: I1206 06:57:35.291862 4945 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:35 crc kubenswrapper[4945]: I1206 06:57:35.313407 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" event={"ID":"957f12af-ad21-4c4a-97a2-ab1b824aa5a0","Type":"ContainerDied","Data":"62a2d51b6a9d7d7381fc1894e81e088920deb46b61702c870bc5daafd1d9d359"} Dec 06 06:57:35 crc kubenswrapper[4945]: I1206 06:57:35.313477 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-456vw" Dec 06 06:57:35 crc kubenswrapper[4945]: I1206 06:57:35.313483 4945 scope.go:117] "RemoveContainer" containerID="58855a5cb5145e6544f569cb20d94290c3da7d7fdae7b3429bfec3d21f8c4a95" Dec 06 06:57:35 crc kubenswrapper[4945]: E1206 06:57:35.935332 4945 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Dec 06 06:57:36 crc kubenswrapper[4945]: I1206 06:57:36.322555 4945 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2b924ff1-dd26-4e47-b90c-144eae88b737" Dec 06 06:57:36 crc kubenswrapper[4945]: I1206 06:57:36.322594 4945 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2b924ff1-dd26-4e47-b90c-144eae88b737" Dec 06 06:57:36 crc kubenswrapper[4945]: I1206 06:57:36.327486 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:36 crc kubenswrapper[4945]: I1206 06:57:36.967613 4945 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d31d16f3-41ca-4592-bd4f-d7c840f28452" Dec 06 06:57:37 crc kubenswrapper[4945]: I1206 06:57:37.329916 4945 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2b924ff1-dd26-4e47-b90c-144eae88b737" Dec 06 06:57:37 crc kubenswrapper[4945]: I1206 06:57:37.330940 4945 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2b924ff1-dd26-4e47-b90c-144eae88b737" Dec 06 06:57:37 crc kubenswrapper[4945]: I1206 06:57:37.333725 4945 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d31d16f3-41ca-4592-bd4f-d7c840f28452" Dec 06 06:57:38 crc kubenswrapper[4945]: I1206 06:57:38.998617 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:57:43 crc kubenswrapper[4945]: I1206 06:57:43.640141 4945 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 06 06:57:43 crc kubenswrapper[4945]: I1206 06:57:43.640509 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 06 06:57:45 crc kubenswrapper[4945]: I1206 06:57:45.543008 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 06:57:45 crc kubenswrapper[4945]: I1206 06:57:45.744740 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 06:57:46 crc kubenswrapper[4945]: I1206 06:57:46.095037 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 06:57:46 crc kubenswrapper[4945]: I1206 06:57:46.875920 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 06:57:46 crc kubenswrapper[4945]: I1206 06:57:46.961511 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:46.970994 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.001517 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.006041 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.135151 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.149156 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.193319 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.555151 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.650624 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.705776 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.896915 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.911034 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 06:57:47 crc kubenswrapper[4945]: I1206 06:57:47.955127 4945 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.127116 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.192666 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.196750 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.247616 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.280096 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.286456 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.593613 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.670962 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.740517 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.928472 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 06:57:48 crc kubenswrapper[4945]: I1206 06:57:48.940100 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.010671 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.034898 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.232446 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.289938 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.355404 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.373509 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.390783 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.423840 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.641039 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.644466 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.660886 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.719107 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.721572 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.910228 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 06:57:49 crc kubenswrapper[4945]: I1206 06:57:49.941413 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 06:57:50 crc kubenswrapper[4945]: I1206 06:57:50.018385 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 06:57:50 crc kubenswrapper[4945]: I1206 06:57:50.177483 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 06:57:50 crc kubenswrapper[4945]: I1206 06:57:50.284012 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 06:57:50 crc kubenswrapper[4945]: I1206 06:57:50.469891 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 06:57:50 crc kubenswrapper[4945]: I1206 06:57:50.558821 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 06:57:50 crc kubenswrapper[4945]: I1206 06:57:50.588623 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 06:57:50 crc kubenswrapper[4945]: I1206 06:57:50.672548 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 06:57:50 crc kubenswrapper[4945]: I1206 06:57:50.838563 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 06:57:50 crc kubenswrapper[4945]: I1206 06:57:50.946409 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 06:57:50 crc kubenswrapper[4945]: I1206 06:57:50.984839 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.012955 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.098547 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.125472 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.131566 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.159895 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.427334 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.428453 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.514259 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.561165 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.568889 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.662058 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.737582 4945 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.936077 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.968321 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 06:57:51 crc kubenswrapper[4945]: I1206 06:57:51.997368 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.003839 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.017112 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.027637 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.064749 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.089775 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.124553 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.203421 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.240982 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.252421 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.423737 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.445975 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.504070 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.509014 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.536037 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.609814 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.647930 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.748706 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.783093 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.818236 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.874337 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.899945 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.951766 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.952237 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 06:57:52 crc kubenswrapper[4945]: I1206 06:57:52.970545 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.050487 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.053340 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.057341 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.160247 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.171611 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.189931 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.278865 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.307535 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.366722 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.398122 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.486114 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.508904 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.537035 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.539659 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.572112 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.605050 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.639723 4945 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.639782 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.639837 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.640404 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"99d1d295024de88c141b9f8d7b99f57b0ce324abf4c7370dd527519050c29df4"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.640545 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://99d1d295024de88c141b9f8d7b99f57b0ce324abf4c7370dd527519050c29df4" gracePeriod=30 Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.763036 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.814602 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 06:57:53 crc kubenswrapper[4945]: I1206 06:57:53.835218 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.005030 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.229969 4945 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.235144 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.237744 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-456vw","openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.237841 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.239402 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.242924 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.263468 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.263447061 podStartE2EDuration="19.263447061s" podCreationTimestamp="2025-12-06 06:57:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:57:54.261780066 +0000 UTC m=+287.716641160" watchObservedRunningTime="2025-12-06 06:57:54.263447061 +0000 UTC m=+287.718308105" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.279221 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.307491 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.343764 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.348249 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.353971 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.478577 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.495574 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.514208 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.548379 4945 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.707032 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.713486 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.759993 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.786377 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.889455 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.960203 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="957f12af-ad21-4c4a-97a2-ab1b824aa5a0" path="/var/lib/kubelet/pods/957f12af-ad21-4c4a-97a2-ab1b824aa5a0/volumes" Dec 06 06:57:54 crc kubenswrapper[4945]: I1206 06:57:54.969118 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:54.992620 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.053863 4945 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.113025 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.173840 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.311443 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.377671 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.412375 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.480015 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.496079 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.496459 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.561632 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.568180 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.577584 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.640179 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.728702 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.825227 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.894130 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.894483 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 06:57:55 crc kubenswrapper[4945]: I1206 06:57:55.911874 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.114946 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.235071 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.239410 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.252489 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.276404 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.327182 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.407751 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.411493 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.492348 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.509520 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.612365 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.644501 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.693034 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.708630 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.730003 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.755182 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.801681 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 06:57:56 crc kubenswrapper[4945]: I1206 06:57:56.856591 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.024733 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.031555 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.117401 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.262477 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.266600 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.270855 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.372767 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.389089 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.502555 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.552517 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.586133 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.612505 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.612516 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.635859 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.639717 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.649455 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.711814 4945 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.712054 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b152fc0d3dfcf1aa5cd5c20554a27524505731efeef856c8069e75a3381eaafe" gracePeriod=5 Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.746960 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.883976 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 06:57:57 crc kubenswrapper[4945]: I1206 06:57:57.918446 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.018058 4945 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.031590 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.062430 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.132788 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.133604 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.187393 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.193414 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.298415 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.299988 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.335916 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.379900 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.380577 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.461428 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.543377 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.549168 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.586871 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.749991 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.815155 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 06:57:58 crc kubenswrapper[4945]: I1206 06:57:58.874362 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 06:57:59 crc kubenswrapper[4945]: I1206 06:57:59.350830 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 06:57:59 crc kubenswrapper[4945]: I1206 06:57:59.533775 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 06:57:59 crc kubenswrapper[4945]: I1206 06:57:59.724038 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 06:57:59 crc kubenswrapper[4945]: I1206 06:57:59.738869 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 06:57:59 crc kubenswrapper[4945]: I1206 06:57:59.788641 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 06:57:59 crc kubenswrapper[4945]: I1206 06:57:59.795896 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 06:57:59 crc kubenswrapper[4945]: I1206 06:57:59.884736 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.006997 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.027718 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.121714 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.468189 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.470439 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.473689 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-79656f7ff7-hl424"] Dec 06 06:58:00 crc kubenswrapper[4945]: E1206 06:58:00.474013 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.474033 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 06:58:00 crc kubenswrapper[4945]: E1206 06:58:00.474045 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" containerName="installer" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.474051 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" containerName="installer" Dec 06 06:58:00 crc kubenswrapper[4945]: E1206 06:58:00.474086 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="957f12af-ad21-4c4a-97a2-ab1b824aa5a0" containerName="oauth-openshift" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.474095 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="957f12af-ad21-4c4a-97a2-ab1b824aa5a0" containerName="oauth-openshift" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.474209 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0af9496b-2588-4cb7-9bc2-922c48b01af2" containerName="installer" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.474260 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="957f12af-ad21-4c4a-97a2-ab1b824aa5a0" containerName="oauth-openshift" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.474272 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.474709 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.483088 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.483181 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.484009 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.484175 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.484177 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.484252 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.484266 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.484182 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.484564 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.484636 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.484670 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.484723 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.488414 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.488688 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79656f7ff7-hl424"] Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.488907 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.507048 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.517496 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.551841 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-template-error\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.551912 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ce539b32-d732-4fb9-b445-b407e0e24bd1-audit-dir\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.551948 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552049 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-template-login\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552102 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552126 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552155 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-audit-policies\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552177 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552197 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552242 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-session\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552267 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7chv\" (UniqueName: \"kubernetes.io/projected/ce539b32-d732-4fb9-b445-b407e0e24bd1-kube-api-access-x7chv\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552316 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-router-certs\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552343 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.552369 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-service-ca\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.591267 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654261 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-router-certs\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654441 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654518 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-service-ca\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654614 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-template-error\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654680 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ce539b32-d732-4fb9-b445-b407e0e24bd1-audit-dir\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654724 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654771 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-template-login\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654811 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654854 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654863 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ce539b32-d732-4fb9-b445-b407e0e24bd1-audit-dir\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.654900 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-audit-policies\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.655796 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.655834 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.655864 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-session\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.655890 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7chv\" (UniqueName: \"kubernetes.io/projected/ce539b32-d732-4fb9-b445-b407e0e24bd1-kube-api-access-x7chv\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.655891 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-service-ca\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.655970 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-audit-policies\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.656415 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.658163 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.662225 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-template-login\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.662682 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.664098 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.664553 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-session\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.664876 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.676575 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-system-router-certs\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.676763 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.679246 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ce539b32-d732-4fb9-b445-b407e0e24bd1-v4-0-config-user-template-error\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.680663 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7chv\" (UniqueName: \"kubernetes.io/projected/ce539b32-d732-4fb9-b445-b407e0e24bd1-kube-api-access-x7chv\") pod \"oauth-openshift-79656f7ff7-hl424\" (UID: \"ce539b32-d732-4fb9-b445-b407e0e24bd1\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.802942 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.821982 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.881470 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 06:58:00 crc kubenswrapper[4945]: I1206 06:58:00.963384 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 06:58:01 crc kubenswrapper[4945]: I1206 06:58:01.006792 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 06:58:01 crc kubenswrapper[4945]: I1206 06:58:01.048074 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 06:58:01 crc kubenswrapper[4945]: I1206 06:58:01.054298 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 06:58:01 crc kubenswrapper[4945]: I1206 06:58:01.070447 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 06:58:01 crc kubenswrapper[4945]: I1206 06:58:01.168703 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 06:58:01 crc kubenswrapper[4945]: I1206 06:58:01.208252 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79656f7ff7-hl424"] Dec 06 06:58:01 crc kubenswrapper[4945]: W1206 06:58:01.214779 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce539b32_d732_4fb9_b445_b407e0e24bd1.slice/crio-e993646ec7806d59677eb8376c37e19389ad755d16dc4ee7311a47da3f1fec2e WatchSource:0}: Error finding container e993646ec7806d59677eb8376c37e19389ad755d16dc4ee7311a47da3f1fec2e: Status 404 returned error can't find the container with id e993646ec7806d59677eb8376c37e19389ad755d16dc4ee7311a47da3f1fec2e Dec 06 06:58:01 crc kubenswrapper[4945]: I1206 06:58:01.478433 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" event={"ID":"ce539b32-d732-4fb9-b445-b407e0e24bd1","Type":"ContainerStarted","Data":"e993646ec7806d59677eb8376c37e19389ad755d16dc4ee7311a47da3f1fec2e"} Dec 06 06:58:01 crc kubenswrapper[4945]: I1206 06:58:01.570824 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 06:58:01 crc kubenswrapper[4945]: I1206 06:58:01.584167 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 06:58:01 crc kubenswrapper[4945]: I1206 06:58:01.787064 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 06:58:02 crc kubenswrapper[4945]: I1206 06:58:02.025136 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 06:58:02 crc kubenswrapper[4945]: I1206 06:58:02.342110 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 06:58:02 crc kubenswrapper[4945]: I1206 06:58:02.401867 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 06:58:02 crc kubenswrapper[4945]: I1206 06:58:02.488734 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" event={"ID":"ce539b32-d732-4fb9-b445-b407e0e24bd1","Type":"ContainerStarted","Data":"fd0f4aab021f228dad99276edfdeb8dbdadb35eb054ce7afbc5859a046cf50bf"} Dec 06 06:58:02 crc kubenswrapper[4945]: I1206 06:58:02.488911 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:02 crc kubenswrapper[4945]: I1206 06:58:02.494725 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" Dec 06 06:58:02 crc kubenswrapper[4945]: I1206 06:58:02.514170 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-79656f7ff7-hl424" podStartSLOduration=54.514150987 podStartE2EDuration="54.514150987s" podCreationTimestamp="2025-12-06 06:57:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:58:02.510666883 +0000 UTC m=+295.965527927" watchObservedRunningTime="2025-12-06 06:58:02.514150987 +0000 UTC m=+295.969012031" Dec 06 06:58:02 crc kubenswrapper[4945]: I1206 06:58:02.860838 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.498258 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.498355 4945 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b152fc0d3dfcf1aa5cd5c20554a27524505731efeef856c8069e75a3381eaafe" exitCode=137 Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.604172 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.604250 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696401 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696462 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696494 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696525 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696566 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696630 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696748 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696788 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696806 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696956 4945 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.696976 4945 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.704678 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.798677 4945 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.798728 4945 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:03 crc kubenswrapper[4945]: I1206 06:58:03.798748 4945 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:04 crc kubenswrapper[4945]: I1206 06:58:04.046501 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 06:58:04 crc kubenswrapper[4945]: I1206 06:58:04.479364 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 06:58:04 crc kubenswrapper[4945]: I1206 06:58:04.508260 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 06:58:04 crc kubenswrapper[4945]: I1206 06:58:04.508498 4945 scope.go:117] "RemoveContainer" containerID="b152fc0d3dfcf1aa5cd5c20554a27524505731efeef856c8069e75a3381eaafe" Dec 06 06:58:04 crc kubenswrapper[4945]: I1206 06:58:04.508548 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 06:58:04 crc kubenswrapper[4945]: I1206 06:58:04.860480 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 06:58:04 crc kubenswrapper[4945]: I1206 06:58:04.961942 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 06 06:58:25 crc kubenswrapper[4945]: I1206 06:58:25.648033 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 06 06:58:25 crc kubenswrapper[4945]: I1206 06:58:25.651568 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 06:58:25 crc kubenswrapper[4945]: I1206 06:58:25.651636 4945 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="99d1d295024de88c141b9f8d7b99f57b0ce324abf4c7370dd527519050c29df4" exitCode=137 Dec 06 06:58:25 crc kubenswrapper[4945]: I1206 06:58:25.651683 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"99d1d295024de88c141b9f8d7b99f57b0ce324abf4c7370dd527519050c29df4"} Dec 06 06:58:25 crc kubenswrapper[4945]: I1206 06:58:25.651761 4945 scope.go:117] "RemoveContainer" containerID="30ef6b0a6440aafac435c036f39bc2ca629f7e88b1f31fd1e1ced83efdba0fe4" Dec 06 06:58:26 crc kubenswrapper[4945]: I1206 06:58:26.663591 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 06 06:58:26 crc kubenswrapper[4945]: I1206 06:58:26.665613 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cb2ea90360558f8cf0e54110ed70bd133c4ce0934e17f99de3e02ef97ba0862e"} Dec 06 06:58:28 crc kubenswrapper[4945]: I1206 06:58:28.998708 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:58:33 crc kubenswrapper[4945]: I1206 06:58:33.639498 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:58:33 crc kubenswrapper[4945]: I1206 06:58:33.644662 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:58:39 crc kubenswrapper[4945]: I1206 06:58:39.003065 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 06:58:54 crc kubenswrapper[4945]: I1206 06:58:54.679771 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x95jj"] Dec 06 06:58:54 crc kubenswrapper[4945]: I1206 06:58:54.681631 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" podUID="e06b2844-0600-4df6-bf5e-749ed15e596a" containerName="controller-manager" containerID="cri-o://1462cbd3aeabfb5e5a0b333916b901cc13097d082dfc56cb4c534f6b40fc7d5b" gracePeriod=30 Dec 06 06:58:54 crc kubenswrapper[4945]: I1206 06:58:54.787617 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp"] Dec 06 06:58:54 crc kubenswrapper[4945]: I1206 06:58:54.787836 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" podUID="dd618c39-c17a-40d9-add1-f87a6f646241" containerName="route-controller-manager" containerID="cri-o://6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031" gracePeriod=30 Dec 06 06:58:54 crc kubenswrapper[4945]: I1206 06:58:54.827198 4945 generic.go:334] "Generic (PLEG): container finished" podID="e06b2844-0600-4df6-bf5e-749ed15e596a" containerID="1462cbd3aeabfb5e5a0b333916b901cc13097d082dfc56cb4c534f6b40fc7d5b" exitCode=0 Dec 06 06:58:54 crc kubenswrapper[4945]: I1206 06:58:54.827664 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" event={"ID":"e06b2844-0600-4df6-bf5e-749ed15e596a","Type":"ContainerDied","Data":"1462cbd3aeabfb5e5a0b333916b901cc13097d082dfc56cb4c534f6b40fc7d5b"} Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.043777 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.147383 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.192027 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e06b2844-0600-4df6-bf5e-749ed15e596a-serving-cert\") pod \"e06b2844-0600-4df6-bf5e-749ed15e596a\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.192097 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-client-ca\") pod \"e06b2844-0600-4df6-bf5e-749ed15e596a\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.192129 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-config\") pod \"e06b2844-0600-4df6-bf5e-749ed15e596a\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.192171 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4sf8\" (UniqueName: \"kubernetes.io/projected/e06b2844-0600-4df6-bf5e-749ed15e596a-kube-api-access-n4sf8\") pod \"e06b2844-0600-4df6-bf5e-749ed15e596a\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.192202 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-proxy-ca-bundles\") pod \"e06b2844-0600-4df6-bf5e-749ed15e596a\" (UID: \"e06b2844-0600-4df6-bf5e-749ed15e596a\") " Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.193203 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e06b2844-0600-4df6-bf5e-749ed15e596a" (UID: "e06b2844-0600-4df6-bf5e-749ed15e596a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.193230 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-client-ca" (OuterVolumeSpecName: "client-ca") pod "e06b2844-0600-4df6-bf5e-749ed15e596a" (UID: "e06b2844-0600-4df6-bf5e-749ed15e596a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.193716 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-config" (OuterVolumeSpecName: "config") pod "e06b2844-0600-4df6-bf5e-749ed15e596a" (UID: "e06b2844-0600-4df6-bf5e-749ed15e596a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.198651 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e06b2844-0600-4df6-bf5e-749ed15e596a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e06b2844-0600-4df6-bf5e-749ed15e596a" (UID: "e06b2844-0600-4df6-bf5e-749ed15e596a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.198651 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e06b2844-0600-4df6-bf5e-749ed15e596a-kube-api-access-n4sf8" (OuterVolumeSpecName: "kube-api-access-n4sf8") pod "e06b2844-0600-4df6-bf5e-749ed15e596a" (UID: "e06b2844-0600-4df6-bf5e-749ed15e596a"). InnerVolumeSpecName "kube-api-access-n4sf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.293678 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-client-ca\") pod \"dd618c39-c17a-40d9-add1-f87a6f646241\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.293756 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkh2m\" (UniqueName: \"kubernetes.io/projected/dd618c39-c17a-40d9-add1-f87a6f646241-kube-api-access-fkh2m\") pod \"dd618c39-c17a-40d9-add1-f87a6f646241\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.293820 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd618c39-c17a-40d9-add1-f87a6f646241-serving-cert\") pod \"dd618c39-c17a-40d9-add1-f87a6f646241\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.293851 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-config\") pod \"dd618c39-c17a-40d9-add1-f87a6f646241\" (UID: \"dd618c39-c17a-40d9-add1-f87a6f646241\") " Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.294078 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.294091 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.294101 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4sf8\" (UniqueName: \"kubernetes.io/projected/e06b2844-0600-4df6-bf5e-749ed15e596a-kube-api-access-n4sf8\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.294112 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e06b2844-0600-4df6-bf5e-749ed15e596a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.294123 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e06b2844-0600-4df6-bf5e-749ed15e596a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.294939 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-config" (OuterVolumeSpecName: "config") pod "dd618c39-c17a-40d9-add1-f87a6f646241" (UID: "dd618c39-c17a-40d9-add1-f87a6f646241"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.295076 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-client-ca" (OuterVolumeSpecName: "client-ca") pod "dd618c39-c17a-40d9-add1-f87a6f646241" (UID: "dd618c39-c17a-40d9-add1-f87a6f646241"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.297812 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd618c39-c17a-40d9-add1-f87a6f646241-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dd618c39-c17a-40d9-add1-f87a6f646241" (UID: "dd618c39-c17a-40d9-add1-f87a6f646241"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.297857 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd618c39-c17a-40d9-add1-f87a6f646241-kube-api-access-fkh2m" (OuterVolumeSpecName: "kube-api-access-fkh2m") pod "dd618c39-c17a-40d9-add1-f87a6f646241" (UID: "dd618c39-c17a-40d9-add1-f87a6f646241"). InnerVolumeSpecName "kube-api-access-fkh2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.394846 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkh2m\" (UniqueName: \"kubernetes.io/projected/dd618c39-c17a-40d9-add1-f87a6f646241-kube-api-access-fkh2m\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.394885 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd618c39-c17a-40d9-add1-f87a6f646241-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.394900 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.394915 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd618c39-c17a-40d9-add1-f87a6f646241-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.833644 4945 generic.go:334] "Generic (PLEG): container finished" podID="dd618c39-c17a-40d9-add1-f87a6f646241" containerID="6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031" exitCode=0 Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.833714 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.833745 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" event={"ID":"dd618c39-c17a-40d9-add1-f87a6f646241","Type":"ContainerDied","Data":"6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031"} Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.835271 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp" event={"ID":"dd618c39-c17a-40d9-add1-f87a6f646241","Type":"ContainerDied","Data":"53266cde319546d78ff1e41e87d82bd34f57bed41a8c25dfdee6b6a15ede6da7"} Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.835320 4945 scope.go:117] "RemoveContainer" containerID="6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.837667 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" event={"ID":"e06b2844-0600-4df6-bf5e-749ed15e596a","Type":"ContainerDied","Data":"cb9c43d89b3347de5f75ba68e4a1cf36a3c36bcdfa7f71d728f266f34ff2b233"} Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.837735 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-x95jj" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.859243 4945 scope.go:117] "RemoveContainer" containerID="6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031" Dec 06 06:58:55 crc kubenswrapper[4945]: E1206 06:58:55.859799 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031\": container with ID starting with 6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031 not found: ID does not exist" containerID="6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.859838 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031"} err="failed to get container status \"6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031\": rpc error: code = NotFound desc = could not find container \"6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031\": container with ID starting with 6585fcc72d611b462a41c5c03210c280ccbbc1e7222061eade044fc21f495031 not found: ID does not exist" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.859865 4945 scope.go:117] "RemoveContainer" containerID="1462cbd3aeabfb5e5a0b333916b901cc13097d082dfc56cb4c534f6b40fc7d5b" Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.862579 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp"] Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.865590 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xrvfp"] Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.872956 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x95jj"] Dec 06 06:58:55 crc kubenswrapper[4945]: I1206 06:58:55.874043 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x95jj"] Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.499396 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl"] Dec 06 06:58:56 crc kubenswrapper[4945]: E1206 06:58:56.499703 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e06b2844-0600-4df6-bf5e-749ed15e596a" containerName="controller-manager" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.499727 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e06b2844-0600-4df6-bf5e-749ed15e596a" containerName="controller-manager" Dec 06 06:58:56 crc kubenswrapper[4945]: E1206 06:58:56.499747 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd618c39-c17a-40d9-add1-f87a6f646241" containerName="route-controller-manager" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.499753 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd618c39-c17a-40d9-add1-f87a6f646241" containerName="route-controller-manager" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.499848 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd618c39-c17a-40d9-add1-f87a6f646241" containerName="route-controller-manager" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.499869 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e06b2844-0600-4df6-bf5e-749ed15e596a" containerName="controller-manager" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.500309 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.502357 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.502593 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f784d6689-qv4s4"] Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.502699 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.503035 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.503088 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.503101 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.503374 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.503412 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.505716 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.505827 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.505962 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.506070 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.506172 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.506240 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.514819 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.518135 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl"] Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.521557 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f784d6689-qv4s4"] Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.609558 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19129773-2406-428b-add4-f17ea13d8ae2-serving-cert\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.609623 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-config\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.609765 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tndw5\" (UniqueName: \"kubernetes.io/projected/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-kube-api-access-tndw5\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.609910 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxs4f\" (UniqueName: \"kubernetes.io/projected/19129773-2406-428b-add4-f17ea13d8ae2-kube-api-access-vxs4f\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.609986 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-serving-cert\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.610049 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-proxy-ca-bundles\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.610092 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-client-ca\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.610113 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-config\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.610150 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-client-ca\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.710840 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxs4f\" (UniqueName: \"kubernetes.io/projected/19129773-2406-428b-add4-f17ea13d8ae2-kube-api-access-vxs4f\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.710891 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-serving-cert\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.710911 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-proxy-ca-bundles\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.710932 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-client-ca\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.710950 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-config\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.710974 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-client-ca\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.711001 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19129773-2406-428b-add4-f17ea13d8ae2-serving-cert\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.711019 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-config\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.711044 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tndw5\" (UniqueName: \"kubernetes.io/projected/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-kube-api-access-tndw5\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.712547 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-config\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.712723 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-client-ca\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.712902 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-client-ca\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.712903 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-proxy-ca-bundles\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.714650 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-config\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.716166 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19129773-2406-428b-add4-f17ea13d8ae2-serving-cert\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.718847 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-serving-cert\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.727102 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tndw5\" (UniqueName: \"kubernetes.io/projected/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-kube-api-access-tndw5\") pod \"route-controller-manager-657648d9fc-n65cl\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.728871 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxs4f\" (UniqueName: \"kubernetes.io/projected/19129773-2406-428b-add4-f17ea13d8ae2-kube-api-access-vxs4f\") pod \"controller-manager-5f784d6689-qv4s4\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.823479 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.835973 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.962570 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd618c39-c17a-40d9-add1-f87a6f646241" path="/var/lib/kubelet/pods/dd618c39-c17a-40d9-add1-f87a6f646241/volumes" Dec 06 06:58:56 crc kubenswrapper[4945]: I1206 06:58:56.963595 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e06b2844-0600-4df6-bf5e-749ed15e596a" path="/var/lib/kubelet/pods/e06b2844-0600-4df6-bf5e-749ed15e596a/volumes" Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.018546 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl"] Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.073974 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f784d6689-qv4s4"] Dec 06 06:58:57 crc kubenswrapper[4945]: W1206 06:58:57.079500 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19129773_2406_428b_add4_f17ea13d8ae2.slice/crio-c2cc8019eda8e72c434223c336de3d7f4606c65a418fb88c94b25fc38fb645d4 WatchSource:0}: Error finding container c2cc8019eda8e72c434223c336de3d7f4606c65a418fb88c94b25fc38fb645d4: Status 404 returned error can't find the container with id c2cc8019eda8e72c434223c336de3d7f4606c65a418fb88c94b25fc38fb645d4 Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.855553 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" event={"ID":"1bbae5a9-8dc6-4269-9c58-b503521ee6ae","Type":"ContainerStarted","Data":"213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b"} Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.856181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" event={"ID":"1bbae5a9-8dc6-4269-9c58-b503521ee6ae","Type":"ContainerStarted","Data":"01043dab0130c2a2fc7df42db8c38deafb3aac6f7bb0c855c87540b14dd4a04b"} Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.856524 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.857631 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" event={"ID":"19129773-2406-428b-add4-f17ea13d8ae2","Type":"ContainerStarted","Data":"e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46"} Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.857877 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.858000 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" event={"ID":"19129773-2406-428b-add4-f17ea13d8ae2","Type":"ContainerStarted","Data":"c2cc8019eda8e72c434223c336de3d7f4606c65a418fb88c94b25fc38fb645d4"} Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.863194 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.868731 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.876216 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" podStartSLOduration=3.876195989 podStartE2EDuration="3.876195989s" podCreationTimestamp="2025-12-06 06:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:58:57.87287103 +0000 UTC m=+351.327732074" watchObservedRunningTime="2025-12-06 06:58:57.876195989 +0000 UTC m=+351.331057033" Dec 06 06:58:57 crc kubenswrapper[4945]: I1206 06:58:57.892134 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" podStartSLOduration=3.892109295 podStartE2EDuration="3.892109295s" podCreationTimestamp="2025-12-06 06:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:58:57.888795886 +0000 UTC m=+351.343656940" watchObservedRunningTime="2025-12-06 06:58:57.892109295 +0000 UTC m=+351.346970349" Dec 06 06:59:08 crc kubenswrapper[4945]: I1206 06:59:08.795802 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:59:08 crc kubenswrapper[4945]: I1206 06:59:08.796685 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:59:14 crc kubenswrapper[4945]: I1206 06:59:14.800897 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl"] Dec 06 06:59:14 crc kubenswrapper[4945]: I1206 06:59:14.801989 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" podUID="1bbae5a9-8dc6-4269-9c58-b503521ee6ae" containerName="route-controller-manager" containerID="cri-o://213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b" gracePeriod=30 Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.813972 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.845113 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj"] Dec 06 06:59:15 crc kubenswrapper[4945]: E1206 06:59:15.845611 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bbae5a9-8dc6-4269-9c58-b503521ee6ae" containerName="route-controller-manager" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.845627 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bbae5a9-8dc6-4269-9c58-b503521ee6ae" containerName="route-controller-manager" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.845755 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bbae5a9-8dc6-4269-9c58-b503521ee6ae" containerName="route-controller-manager" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.846308 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.858807 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj"] Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.904225 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-client-ca\") pod \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.904653 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tndw5\" (UniqueName: \"kubernetes.io/projected/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-kube-api-access-tndw5\") pod \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.904713 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-serving-cert\") pod \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.904787 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-config\") pod \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\" (UID: \"1bbae5a9-8dc6-4269-9c58-b503521ee6ae\") " Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.905078 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c1f436-d21b-48ec-a6d8-d95efa793043-config\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.905153 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0c1f436-d21b-48ec-a6d8-d95efa793043-client-ca\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.905294 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0c1f436-d21b-48ec-a6d8-d95efa793043-serving-cert\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.905377 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w2hj\" (UniqueName: \"kubernetes.io/projected/c0c1f436-d21b-48ec-a6d8-d95efa793043-kube-api-access-8w2hj\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.905637 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-config" (OuterVolumeSpecName: "config") pod "1bbae5a9-8dc6-4269-9c58-b503521ee6ae" (UID: "1bbae5a9-8dc6-4269-9c58-b503521ee6ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.905761 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-client-ca" (OuterVolumeSpecName: "client-ca") pod "1bbae5a9-8dc6-4269-9c58-b503521ee6ae" (UID: "1bbae5a9-8dc6-4269-9c58-b503521ee6ae"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.911417 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bbae5a9-8dc6-4269-9c58-b503521ee6ae" (UID: "1bbae5a9-8dc6-4269-9c58-b503521ee6ae"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.911799 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-kube-api-access-tndw5" (OuterVolumeSpecName: "kube-api-access-tndw5") pod "1bbae5a9-8dc6-4269-9c58-b503521ee6ae" (UID: "1bbae5a9-8dc6-4269-9c58-b503521ee6ae"). InnerVolumeSpecName "kube-api-access-tndw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.961138 4945 generic.go:334] "Generic (PLEG): container finished" podID="1bbae5a9-8dc6-4269-9c58-b503521ee6ae" containerID="213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b" exitCode=0 Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.961238 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" event={"ID":"1bbae5a9-8dc6-4269-9c58-b503521ee6ae","Type":"ContainerDied","Data":"213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b"} Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.961693 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" event={"ID":"1bbae5a9-8dc6-4269-9c58-b503521ee6ae","Type":"ContainerDied","Data":"01043dab0130c2a2fc7df42db8c38deafb3aac6f7bb0c855c87540b14dd4a04b"} Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.961815 4945 scope.go:117] "RemoveContainer" containerID="213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.961298 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.981729 4945 scope.go:117] "RemoveContainer" containerID="213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b" Dec 06 06:59:15 crc kubenswrapper[4945]: E1206 06:59:15.982341 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b\": container with ID starting with 213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b not found: ID does not exist" containerID="213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.982383 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b"} err="failed to get container status \"213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b\": rpc error: code = NotFound desc = could not find container \"213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b\": container with ID starting with 213479d9b7a5e151a5c490b1c404ba31fb50b7c18fff9cb1b97c00b4acc9655b not found: ID does not exist" Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.992322 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl"] Dec 06 06:59:15 crc kubenswrapper[4945]: I1206 06:59:15.995565 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657648d9fc-n65cl"] Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.006844 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0c1f436-d21b-48ec-a6d8-d95efa793043-client-ca\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.006909 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0c1f436-d21b-48ec-a6d8-d95efa793043-serving-cert\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.006932 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w2hj\" (UniqueName: \"kubernetes.io/projected/c0c1f436-d21b-48ec-a6d8-d95efa793043-kube-api-access-8w2hj\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.006979 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c1f436-d21b-48ec-a6d8-d95efa793043-config\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.007041 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.007057 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tndw5\" (UniqueName: \"kubernetes.io/projected/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-kube-api-access-tndw5\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.007188 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.007234 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bbae5a9-8dc6-4269-9c58-b503521ee6ae-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.009130 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0c1f436-d21b-48ec-a6d8-d95efa793043-client-ca\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.009165 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c1f436-d21b-48ec-a6d8-d95efa793043-config\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.010904 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0c1f436-d21b-48ec-a6d8-d95efa793043-serving-cert\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.029088 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w2hj\" (UniqueName: \"kubernetes.io/projected/c0c1f436-d21b-48ec-a6d8-d95efa793043-kube-api-access-8w2hj\") pod \"route-controller-manager-7599bb98c7-r56mj\" (UID: \"c0c1f436-d21b-48ec-a6d8-d95efa793043\") " pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.164356 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.581957 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj"] Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.961589 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bbae5a9-8dc6-4269-9c58-b503521ee6ae" path="/var/lib/kubelet/pods/1bbae5a9-8dc6-4269-9c58-b503521ee6ae/volumes" Dec 06 06:59:16 crc kubenswrapper[4945]: I1206 06:59:16.968506 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" event={"ID":"c0c1f436-d21b-48ec-a6d8-d95efa793043","Type":"ContainerStarted","Data":"36d0b60fcf84c7347944f9fb5a145de680cb23b7441524b9f110fced781e6703"} Dec 06 06:59:17 crc kubenswrapper[4945]: I1206 06:59:17.976009 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" event={"ID":"c0c1f436-d21b-48ec-a6d8-d95efa793043","Type":"ContainerStarted","Data":"38b47c767b7e8f8220275f52515b651a2cabc05fa65efa97111088c610e1df61"} Dec 06 06:59:18 crc kubenswrapper[4945]: I1206 06:59:18.983209 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:18 crc kubenswrapper[4945]: I1206 06:59:18.989352 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" Dec 06 06:59:19 crc kubenswrapper[4945]: I1206 06:59:19.009143 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7599bb98c7-r56mj" podStartSLOduration=5.00911935 podStartE2EDuration="5.00911935s" podCreationTimestamp="2025-12-06 06:59:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:59:19.002876713 +0000 UTC m=+372.457737757" watchObservedRunningTime="2025-12-06 06:59:19.00911935 +0000 UTC m=+372.463980384" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.004993 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lslh4"] Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.006506 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.030072 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lslh4"] Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.054724 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.054839 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxt2s\" (UniqueName: \"kubernetes.io/projected/5531ea8e-2553-4b9d-a950-e8dfefca36ff-kube-api-access-jxt2s\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.054991 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5531ea8e-2553-4b9d-a950-e8dfefca36ff-registry-tls\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.055046 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5531ea8e-2553-4b9d-a950-e8dfefca36ff-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.055091 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5531ea8e-2553-4b9d-a950-e8dfefca36ff-trusted-ca\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.055368 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5531ea8e-2553-4b9d-a950-e8dfefca36ff-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.055489 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5531ea8e-2553-4b9d-a950-e8dfefca36ff-bound-sa-token\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.056105 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5531ea8e-2553-4b9d-a950-e8dfefca36ff-registry-certificates\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.077718 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.157635 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxt2s\" (UniqueName: \"kubernetes.io/projected/5531ea8e-2553-4b9d-a950-e8dfefca36ff-kube-api-access-jxt2s\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.157709 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5531ea8e-2553-4b9d-a950-e8dfefca36ff-registry-tls\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.157743 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5531ea8e-2553-4b9d-a950-e8dfefca36ff-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.157771 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5531ea8e-2553-4b9d-a950-e8dfefca36ff-trusted-ca\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.157809 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5531ea8e-2553-4b9d-a950-e8dfefca36ff-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.157850 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5531ea8e-2553-4b9d-a950-e8dfefca36ff-bound-sa-token\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.157890 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5531ea8e-2553-4b9d-a950-e8dfefca36ff-registry-certificates\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.158522 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5531ea8e-2553-4b9d-a950-e8dfefca36ff-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.159837 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5531ea8e-2553-4b9d-a950-e8dfefca36ff-trusted-ca\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.159924 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5531ea8e-2553-4b9d-a950-e8dfefca36ff-registry-certificates\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.165982 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5531ea8e-2553-4b9d-a950-e8dfefca36ff-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.166038 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5531ea8e-2553-4b9d-a950-e8dfefca36ff-registry-tls\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.176470 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5531ea8e-2553-4b9d-a950-e8dfefca36ff-bound-sa-token\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.178627 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxt2s\" (UniqueName: \"kubernetes.io/projected/5531ea8e-2553-4b9d-a950-e8dfefca36ff-kube-api-access-jxt2s\") pod \"image-registry-66df7c8f76-lslh4\" (UID: \"5531ea8e-2553-4b9d-a950-e8dfefca36ff\") " pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.325062 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.673764 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5f784d6689-qv4s4"] Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.674145 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" podUID="19129773-2406-428b-add4-f17ea13d8ae2" containerName="controller-manager" containerID="cri-o://e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46" gracePeriod=30 Dec 06 06:59:34 crc kubenswrapper[4945]: I1206 06:59:34.728095 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lslh4"] Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.043122 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.072712 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" event={"ID":"5531ea8e-2553-4b9d-a950-e8dfefca36ff","Type":"ContainerStarted","Data":"7b1c7052e04bb0811b5eccc832d4dbf1a62df05cab4d4ab36577cd7eacce8284"} Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.072786 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" event={"ID":"5531ea8e-2553-4b9d-a950-e8dfefca36ff","Type":"ContainerStarted","Data":"76b0c19f4e6cc7ba6b70b27d723c92b02b575d432a2c1aa9540ebb976d396d91"} Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.073815 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.075497 4945 generic.go:334] "Generic (PLEG): container finished" podID="19129773-2406-428b-add4-f17ea13d8ae2" containerID="e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46" exitCode=0 Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.075525 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" event={"ID":"19129773-2406-428b-add4-f17ea13d8ae2","Type":"ContainerDied","Data":"e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46"} Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.075540 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" event={"ID":"19129773-2406-428b-add4-f17ea13d8ae2","Type":"ContainerDied","Data":"c2cc8019eda8e72c434223c336de3d7f4606c65a418fb88c94b25fc38fb645d4"} Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.075559 4945 scope.go:117] "RemoveContainer" containerID="e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.075653 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f784d6689-qv4s4" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.081363 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-proxy-ca-bundles\") pod \"19129773-2406-428b-add4-f17ea13d8ae2\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.081396 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-client-ca\") pod \"19129773-2406-428b-add4-f17ea13d8ae2\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.081454 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-config\") pod \"19129773-2406-428b-add4-f17ea13d8ae2\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.081480 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19129773-2406-428b-add4-f17ea13d8ae2-serving-cert\") pod \"19129773-2406-428b-add4-f17ea13d8ae2\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.081552 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxs4f\" (UniqueName: \"kubernetes.io/projected/19129773-2406-428b-add4-f17ea13d8ae2-kube-api-access-vxs4f\") pod \"19129773-2406-428b-add4-f17ea13d8ae2\" (UID: \"19129773-2406-428b-add4-f17ea13d8ae2\") " Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.082149 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-client-ca" (OuterVolumeSpecName: "client-ca") pod "19129773-2406-428b-add4-f17ea13d8ae2" (UID: "19129773-2406-428b-add4-f17ea13d8ae2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.082471 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "19129773-2406-428b-add4-f17ea13d8ae2" (UID: "19129773-2406-428b-add4-f17ea13d8ae2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.083225 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-config" (OuterVolumeSpecName: "config") pod "19129773-2406-428b-add4-f17ea13d8ae2" (UID: "19129773-2406-428b-add4-f17ea13d8ae2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.093653 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19129773-2406-428b-add4-f17ea13d8ae2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "19129773-2406-428b-add4-f17ea13d8ae2" (UID: "19129773-2406-428b-add4-f17ea13d8ae2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.096815 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19129773-2406-428b-add4-f17ea13d8ae2-kube-api-access-vxs4f" (OuterVolumeSpecName: "kube-api-access-vxs4f") pod "19129773-2406-428b-add4-f17ea13d8ae2" (UID: "19129773-2406-428b-add4-f17ea13d8ae2"). InnerVolumeSpecName "kube-api-access-vxs4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.112045 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" podStartSLOduration=2.112020796 podStartE2EDuration="2.112020796s" podCreationTimestamp="2025-12-06 06:59:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:59:35.10434216 +0000 UTC m=+388.559203234" watchObservedRunningTime="2025-12-06 06:59:35.112020796 +0000 UTC m=+388.566881850" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.135087 4945 scope.go:117] "RemoveContainer" containerID="e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46" Dec 06 06:59:35 crc kubenswrapper[4945]: E1206 06:59:35.136109 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46\": container with ID starting with e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46 not found: ID does not exist" containerID="e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.136344 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46"} err="failed to get container status \"e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46\": rpc error: code = NotFound desc = could not find container \"e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46\": container with ID starting with e82b1f47a8bef361a3ac2b1db00a7b34bd34e1362145ef9e27547390e9b1da46 not found: ID does not exist" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.183863 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxs4f\" (UniqueName: \"kubernetes.io/projected/19129773-2406-428b-add4-f17ea13d8ae2-kube-api-access-vxs4f\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.183907 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.183922 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.183937 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19129773-2406-428b-add4-f17ea13d8ae2-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.183949 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19129773-2406-428b-add4-f17ea13d8ae2-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.410627 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5f784d6689-qv4s4"] Dec 06 06:59:35 crc kubenswrapper[4945]: I1206 06:59:35.413770 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5f784d6689-qv4s4"] Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.528045 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-88f7fd697-zspqr"] Dec 06 06:59:36 crc kubenswrapper[4945]: E1206 06:59:36.528388 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19129773-2406-428b-add4-f17ea13d8ae2" containerName="controller-manager" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.528409 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="19129773-2406-428b-add4-f17ea13d8ae2" containerName="controller-manager" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.528541 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="19129773-2406-428b-add4-f17ea13d8ae2" containerName="controller-manager" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.529145 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.533368 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.535352 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.535670 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.535840 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.541219 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.541511 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.544565 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.546783 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-88f7fd697-zspqr"] Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.603843 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53baa70d-c4ad-4269-b826-9368986243f6-serving-cert\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.603906 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njb4l\" (UniqueName: \"kubernetes.io/projected/53baa70d-c4ad-4269-b826-9368986243f6-kube-api-access-njb4l\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.604061 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53baa70d-c4ad-4269-b826-9368986243f6-proxy-ca-bundles\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.604244 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53baa70d-c4ad-4269-b826-9368986243f6-client-ca\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.604300 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53baa70d-c4ad-4269-b826-9368986243f6-config\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.706015 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53baa70d-c4ad-4269-b826-9368986243f6-serving-cert\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.706522 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njb4l\" (UniqueName: \"kubernetes.io/projected/53baa70d-c4ad-4269-b826-9368986243f6-kube-api-access-njb4l\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.706592 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53baa70d-c4ad-4269-b826-9368986243f6-proxy-ca-bundles\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.706681 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53baa70d-c4ad-4269-b826-9368986243f6-client-ca\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.706712 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53baa70d-c4ad-4269-b826-9368986243f6-config\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.708142 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53baa70d-c4ad-4269-b826-9368986243f6-config\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.708200 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53baa70d-c4ad-4269-b826-9368986243f6-client-ca\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.708273 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53baa70d-c4ad-4269-b826-9368986243f6-proxy-ca-bundles\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.714827 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53baa70d-c4ad-4269-b826-9368986243f6-serving-cert\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.725900 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njb4l\" (UniqueName: \"kubernetes.io/projected/53baa70d-c4ad-4269-b826-9368986243f6-kube-api-access-njb4l\") pod \"controller-manager-88f7fd697-zspqr\" (UID: \"53baa70d-c4ad-4269-b826-9368986243f6\") " pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.861662 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:36 crc kubenswrapper[4945]: I1206 06:59:36.966968 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19129773-2406-428b-add4-f17ea13d8ae2" path="/var/lib/kubelet/pods/19129773-2406-428b-add4-f17ea13d8ae2/volumes" Dec 06 06:59:37 crc kubenswrapper[4945]: I1206 06:59:37.075203 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-88f7fd697-zspqr"] Dec 06 06:59:38 crc kubenswrapper[4945]: I1206 06:59:38.099970 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" event={"ID":"53baa70d-c4ad-4269-b826-9368986243f6","Type":"ContainerStarted","Data":"d462b1c72a101a4fb82dda333a712dba5ec64fcdcdf83da87deb64bafb4bbd92"} Dec 06 06:59:38 crc kubenswrapper[4945]: I1206 06:59:38.100695 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" event={"ID":"53baa70d-c4ad-4269-b826-9368986243f6","Type":"ContainerStarted","Data":"d931d5cbdffe41fffc9a3fc489d83fae4287849f49cf908ee7bddb552c6e4256"} Dec 06 06:59:38 crc kubenswrapper[4945]: I1206 06:59:38.124662 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" podStartSLOduration=4.124634761 podStartE2EDuration="4.124634761s" podCreationTimestamp="2025-12-06 06:59:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:59:38.122092753 +0000 UTC m=+391.576953807" watchObservedRunningTime="2025-12-06 06:59:38.124634761 +0000 UTC m=+391.579495805" Dec 06 06:59:38 crc kubenswrapper[4945]: I1206 06:59:38.795630 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:59:38 crc kubenswrapper[4945]: I1206 06:59:38.795713 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:59:39 crc kubenswrapper[4945]: I1206 06:59:39.105758 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:39 crc kubenswrapper[4945]: I1206 06:59:39.112235 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-88f7fd697-zspqr" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.247836 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-csg4k"] Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.249146 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-csg4k" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" containerName="registry-server" containerID="cri-o://b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11" gracePeriod=30 Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.254276 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xsktd"] Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.257758 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xsktd" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerName="registry-server" containerID="cri-o://414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64" gracePeriod=30 Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.271029 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8skx"] Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.271344 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" podUID="8d7466cb-add7-436c-8641-0bf1bda1afbd" containerName="marketplace-operator" containerID="cri-o://aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17" gracePeriod=30 Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.286799 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5w5w9"] Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.287218 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5w5w9" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerName="registry-server" containerID="cri-o://cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f" gracePeriod=30 Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.298702 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fvjzs"] Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.299077 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fvjzs" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerName="registry-server" containerID="cri-o://a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0" gracePeriod=30 Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.309516 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6rgp5"] Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.310646 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.335253 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6rgp5"] Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.445214 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrk8x\" (UniqueName: \"kubernetes.io/projected/5d136274-fa83-4411-845d-5d5cc3cff1b7-kube-api-access-zrk8x\") pod \"marketplace-operator-79b997595-6rgp5\" (UID: \"5d136274-fa83-4411-845d-5d5cc3cff1b7\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.445709 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d136274-fa83-4411-845d-5d5cc3cff1b7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6rgp5\" (UID: \"5d136274-fa83-4411-845d-5d5cc3cff1b7\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.445736 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d136274-fa83-4411-845d-5d5cc3cff1b7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6rgp5\" (UID: \"5d136274-fa83-4411-845d-5d5cc3cff1b7\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.546970 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrk8x\" (UniqueName: \"kubernetes.io/projected/5d136274-fa83-4411-845d-5d5cc3cff1b7-kube-api-access-zrk8x\") pod \"marketplace-operator-79b997595-6rgp5\" (UID: \"5d136274-fa83-4411-845d-5d5cc3cff1b7\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.547050 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d136274-fa83-4411-845d-5d5cc3cff1b7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6rgp5\" (UID: \"5d136274-fa83-4411-845d-5d5cc3cff1b7\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.547072 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d136274-fa83-4411-845d-5d5cc3cff1b7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6rgp5\" (UID: \"5d136274-fa83-4411-845d-5d5cc3cff1b7\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.548944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d136274-fa83-4411-845d-5d5cc3cff1b7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6rgp5\" (UID: \"5d136274-fa83-4411-845d-5d5cc3cff1b7\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.564047 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d136274-fa83-4411-845d-5d5cc3cff1b7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6rgp5\" (UID: \"5d136274-fa83-4411-845d-5d5cc3cff1b7\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.574730 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrk8x\" (UniqueName: \"kubernetes.io/projected/5d136274-fa83-4411-845d-5d5cc3cff1b7-kube-api-access-zrk8x\") pod \"marketplace-operator-79b997595-6rgp5\" (UID: \"5d136274-fa83-4411-845d-5d5cc3cff1b7\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.648530 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:43 crc kubenswrapper[4945]: I1206 06:59:43.966583 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.016862 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.019166 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.054131 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.063223 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-operator-metrics\") pod \"8d7466cb-add7-436c-8641-0bf1bda1afbd\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.063352 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-utilities\") pod \"fb6b366b-20cd-404c-be4c-b07df65de3bb\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.063409 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kd96d\" (UniqueName: \"kubernetes.io/projected/fb6b366b-20cd-404c-be4c-b07df65de3bb-kube-api-access-kd96d\") pod \"fb6b366b-20cd-404c-be4c-b07df65de3bb\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.063454 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-trusted-ca\") pod \"8d7466cb-add7-436c-8641-0bf1bda1afbd\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.063475 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4k4wt\" (UniqueName: \"kubernetes.io/projected/8d7466cb-add7-436c-8641-0bf1bda1afbd-kube-api-access-4k4wt\") pod \"8d7466cb-add7-436c-8641-0bf1bda1afbd\" (UID: \"8d7466cb-add7-436c-8641-0bf1bda1afbd\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.063515 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-catalog-content\") pod \"fb6b366b-20cd-404c-be4c-b07df65de3bb\" (UID: \"fb6b366b-20cd-404c-be4c-b07df65de3bb\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.063562 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfw44\" (UniqueName: \"kubernetes.io/projected/319506ca-cc5d-424b-a84f-e7fab30c8a76-kube-api-access-kfw44\") pod \"319506ca-cc5d-424b-a84f-e7fab30c8a76\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.063591 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-catalog-content\") pod \"319506ca-cc5d-424b-a84f-e7fab30c8a76\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.063662 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-utilities\") pod \"319506ca-cc5d-424b-a84f-e7fab30c8a76\" (UID: \"319506ca-cc5d-424b-a84f-e7fab30c8a76\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.065967 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-utilities" (OuterVolumeSpecName: "utilities") pod "fb6b366b-20cd-404c-be4c-b07df65de3bb" (UID: "fb6b366b-20cd-404c-be4c-b07df65de3bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.066331 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-utilities" (OuterVolumeSpecName: "utilities") pod "319506ca-cc5d-424b-a84f-e7fab30c8a76" (UID: "319506ca-cc5d-424b-a84f-e7fab30c8a76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.067504 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "8d7466cb-add7-436c-8641-0bf1bda1afbd" (UID: "8d7466cb-add7-436c-8641-0bf1bda1afbd"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.069909 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.076990 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/319506ca-cc5d-424b-a84f-e7fab30c8a76-kube-api-access-kfw44" (OuterVolumeSpecName: "kube-api-access-kfw44") pod "319506ca-cc5d-424b-a84f-e7fab30c8a76" (UID: "319506ca-cc5d-424b-a84f-e7fab30c8a76"). InnerVolumeSpecName "kube-api-access-kfw44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.077273 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "8d7466cb-add7-436c-8641-0bf1bda1afbd" (UID: "8d7466cb-add7-436c-8641-0bf1bda1afbd"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.081218 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7466cb-add7-436c-8641-0bf1bda1afbd-kube-api-access-4k4wt" (OuterVolumeSpecName: "kube-api-access-4k4wt") pod "8d7466cb-add7-436c-8641-0bf1bda1afbd" (UID: "8d7466cb-add7-436c-8641-0bf1bda1afbd"). InnerVolumeSpecName "kube-api-access-4k4wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.081434 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb6b366b-20cd-404c-be4c-b07df65de3bb-kube-api-access-kd96d" (OuterVolumeSpecName: "kube-api-access-kd96d") pod "fb6b366b-20cd-404c-be4c-b07df65de3bb" (UID: "fb6b366b-20cd-404c-be4c-b07df65de3bb"). InnerVolumeSpecName "kube-api-access-kd96d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.133077 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb6b366b-20cd-404c-be4c-b07df65de3bb" (UID: "fb6b366b-20cd-404c-be4c-b07df65de3bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.148525 4945 generic.go:334] "Generic (PLEG): container finished" podID="8d7466cb-add7-436c-8641-0bf1bda1afbd" containerID="aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17" exitCode=0 Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.148596 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" event={"ID":"8d7466cb-add7-436c-8641-0bf1bda1afbd","Type":"ContainerDied","Data":"aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17"} Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.148629 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.148669 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r8skx" event={"ID":"8d7466cb-add7-436c-8641-0bf1bda1afbd","Type":"ContainerDied","Data":"a270ceb9f171a5167da5f367a842e4ae81ac5fb2ff9ab940f7cb7bbd60dbce8f"} Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.148697 4945 scope.go:117] "RemoveContainer" containerID="aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.156315 4945 generic.go:334] "Generic (PLEG): container finished" podID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerID="414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64" exitCode=0 Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.156412 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsktd" event={"ID":"fb6b366b-20cd-404c-be4c-b07df65de3bb","Type":"ContainerDied","Data":"414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64"} Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.156456 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsktd" event={"ID":"fb6b366b-20cd-404c-be4c-b07df65de3bb","Type":"ContainerDied","Data":"7ffa8fe26638284cfb1811dbe235580ab403321bf83be87a250da3e7e8de6fbc"} Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.156630 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsktd" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.164830 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq8mz\" (UniqueName: \"kubernetes.io/projected/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-kube-api-access-bq8mz\") pod \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.164933 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcmfd\" (UniqueName: \"kubernetes.io/projected/e315043e-3530-4af3-992b-8647e92bfa2d-kube-api-access-xcmfd\") pod \"e315043e-3530-4af3-992b-8647e92bfa2d\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.164982 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-utilities\") pod \"e315043e-3530-4af3-992b-8647e92bfa2d\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165010 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-catalog-content\") pod \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165070 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-utilities\") pod \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\" (UID: \"4c50068a-69e4-4063-8f08-e4dc4e51fc8d\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165099 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-catalog-content\") pod \"e315043e-3530-4af3-992b-8647e92bfa2d\" (UID: \"e315043e-3530-4af3-992b-8647e92bfa2d\") " Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165436 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165457 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kd96d\" (UniqueName: \"kubernetes.io/projected/fb6b366b-20cd-404c-be4c-b07df65de3bb-kube-api-access-kd96d\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165471 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4k4wt\" (UniqueName: \"kubernetes.io/projected/8d7466cb-add7-436c-8641-0bf1bda1afbd-kube-api-access-4k4wt\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165488 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165501 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb6b366b-20cd-404c-be4c-b07df65de3bb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165513 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfw44\" (UniqueName: \"kubernetes.io/projected/319506ca-cc5d-424b-a84f-e7fab30c8a76-kube-api-access-kfw44\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165526 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.165542 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d7466cb-add7-436c-8641-0bf1bda1afbd-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.166660 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-utilities" (OuterVolumeSpecName: "utilities") pod "e315043e-3530-4af3-992b-8647e92bfa2d" (UID: "e315043e-3530-4af3-992b-8647e92bfa2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.167436 4945 generic.go:334] "Generic (PLEG): container finished" podID="e315043e-3530-4af3-992b-8647e92bfa2d" containerID="b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11" exitCode=0 Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.167654 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csg4k" event={"ID":"e315043e-3530-4af3-992b-8647e92bfa2d","Type":"ContainerDied","Data":"b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11"} Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.167750 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csg4k" event={"ID":"e315043e-3530-4af3-992b-8647e92bfa2d","Type":"ContainerDied","Data":"3643b4b1d32edf9c9db7bd456deba388815b4d99f427f9e11e074037d7163987"} Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.167898 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csg4k" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.169900 4945 scope.go:117] "RemoveContainer" containerID="aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.170853 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17\": container with ID starting with aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17 not found: ID does not exist" containerID="aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.170906 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17"} err="failed to get container status \"aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17\": rpc error: code = NotFound desc = could not find container \"aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17\": container with ID starting with aea12e2ae9aa9e8bd4df027d9d2e45f54b2349c2674ce8875bff3152c0128c17 not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.170942 4945 scope.go:117] "RemoveContainer" containerID="414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.171259 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e315043e-3530-4af3-992b-8647e92bfa2d-kube-api-access-xcmfd" (OuterVolumeSpecName: "kube-api-access-xcmfd") pod "e315043e-3530-4af3-992b-8647e92bfa2d" (UID: "e315043e-3530-4af3-992b-8647e92bfa2d"). InnerVolumeSpecName "kube-api-access-xcmfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.171423 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-utilities" (OuterVolumeSpecName: "utilities") pod "4c50068a-69e4-4063-8f08-e4dc4e51fc8d" (UID: "4c50068a-69e4-4063-8f08-e4dc4e51fc8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.173460 4945 generic.go:334] "Generic (PLEG): container finished" podID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerID="cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f" exitCode=0 Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.173522 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w5w9" event={"ID":"4c50068a-69e4-4063-8f08-e4dc4e51fc8d","Type":"ContainerDied","Data":"cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f"} Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.173549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w5w9" event={"ID":"4c50068a-69e4-4063-8f08-e4dc4e51fc8d","Type":"ContainerDied","Data":"071f22cbc7765303f63617f236b07b4ddb13c6b8f96fd788658a0670f2c22992"} Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.173608 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5w5w9" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.174330 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-kube-api-access-bq8mz" (OuterVolumeSpecName: "kube-api-access-bq8mz") pod "4c50068a-69e4-4063-8f08-e4dc4e51fc8d" (UID: "4c50068a-69e4-4063-8f08-e4dc4e51fc8d"). InnerVolumeSpecName "kube-api-access-bq8mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.180536 4945 generic.go:334] "Generic (PLEG): container finished" podID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerID="a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0" exitCode=0 Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.180732 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvjzs" event={"ID":"319506ca-cc5d-424b-a84f-e7fab30c8a76","Type":"ContainerDied","Data":"a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0"} Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.180842 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvjzs" event={"ID":"319506ca-cc5d-424b-a84f-e7fab30c8a76","Type":"ContainerDied","Data":"dc4897c7f8eb297444fa4f2bf2865d99788fc72346bbff9f3cfd2c84ecdb337e"} Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.181000 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fvjzs" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.199106 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8skx"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.203145 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8skx"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.206857 4945 scope.go:117] "RemoveContainer" containerID="ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.213168 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c50068a-69e4-4063-8f08-e4dc4e51fc8d" (UID: "4c50068a-69e4-4063-8f08-e4dc4e51fc8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.215674 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xsktd"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.218962 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xsktd"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.232627 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "319506ca-cc5d-424b-a84f-e7fab30c8a76" (UID: "319506ca-cc5d-424b-a84f-e7fab30c8a76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.240878 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6rgp5"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.242866 4945 scope.go:117] "RemoveContainer" containerID="bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.249488 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e315043e-3530-4af3-992b-8647e92bfa2d" (UID: "e315043e-3530-4af3-992b-8647e92bfa2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.267544 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319506ca-cc5d-424b-a84f-e7fab30c8a76-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.267601 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq8mz\" (UniqueName: \"kubernetes.io/projected/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-kube-api-access-bq8mz\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.267620 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcmfd\" (UniqueName: \"kubernetes.io/projected/e315043e-3530-4af3-992b-8647e92bfa2d-kube-api-access-xcmfd\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.267636 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.267648 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.267663 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c50068a-69e4-4063-8f08-e4dc4e51fc8d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.267676 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e315043e-3530-4af3-992b-8647e92bfa2d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.277598 4945 scope.go:117] "RemoveContainer" containerID="414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.278203 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64\": container with ID starting with 414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64 not found: ID does not exist" containerID="414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.278242 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64"} err="failed to get container status \"414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64\": rpc error: code = NotFound desc = could not find container \"414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64\": container with ID starting with 414bf0cfbc4f3353fb69326f1d740315547ce217cb9b65b9bd0ded0b4470cc64 not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.278374 4945 scope.go:117] "RemoveContainer" containerID="ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.279370 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf\": container with ID starting with ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf not found: ID does not exist" containerID="ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.279443 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf"} err="failed to get container status \"ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf\": rpc error: code = NotFound desc = could not find container \"ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf\": container with ID starting with ba90c3a844bd16e43f459b5e20d9b16eca523fda62ef3069447970181650f5cf not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.279466 4945 scope.go:117] "RemoveContainer" containerID="bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.280296 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914\": container with ID starting with bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914 not found: ID does not exist" containerID="bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.280454 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914"} err="failed to get container status \"bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914\": rpc error: code = NotFound desc = could not find container \"bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914\": container with ID starting with bd1a78d443b43e7a1b452440cacab5ac8254e3f85903c1adee4607edbfc1c914 not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.280535 4945 scope.go:117] "RemoveContainer" containerID="b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.312157 4945 scope.go:117] "RemoveContainer" containerID="1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.341987 4945 scope.go:117] "RemoveContainer" containerID="00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.361640 4945 scope.go:117] "RemoveContainer" containerID="b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.362239 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11\": container with ID starting with b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11 not found: ID does not exist" containerID="b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.362300 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11"} err="failed to get container status \"b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11\": rpc error: code = NotFound desc = could not find container \"b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11\": container with ID starting with b89e979d4180d442a14ef28c0481f5c210d2fdace6ce7514e27dc7d70c234e11 not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.362338 4945 scope.go:117] "RemoveContainer" containerID="1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.363048 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6\": container with ID starting with 1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6 not found: ID does not exist" containerID="1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.363101 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6"} err="failed to get container status \"1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6\": rpc error: code = NotFound desc = could not find container \"1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6\": container with ID starting with 1b054b75c3a1553ec0a0ddc83cf3b0b2556806a644c7d222b88f61f3d5c0c4f6 not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.363139 4945 scope.go:117] "RemoveContainer" containerID="00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.363565 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd\": container with ID starting with 00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd not found: ID does not exist" containerID="00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.363589 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd"} err="failed to get container status \"00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd\": rpc error: code = NotFound desc = could not find container \"00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd\": container with ID starting with 00d4fc974ae39c5cc06d862e256f0b060d20fabb10fc7edc4af7460df40449dd not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.363604 4945 scope.go:117] "RemoveContainer" containerID="cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.382520 4945 scope.go:117] "RemoveContainer" containerID="2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.401146 4945 scope.go:117] "RemoveContainer" containerID="534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.431414 4945 scope.go:117] "RemoveContainer" containerID="cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.432992 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f\": container with ID starting with cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f not found: ID does not exist" containerID="cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.433041 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f"} err="failed to get container status \"cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f\": rpc error: code = NotFound desc = could not find container \"cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f\": container with ID starting with cedea8a386a67a54b26e47a7c729a81e8ffc7f9f42a9c373f7cf001485b5615f not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.433078 4945 scope.go:117] "RemoveContainer" containerID="2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.433614 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37\": container with ID starting with 2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37 not found: ID does not exist" containerID="2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.433635 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37"} err="failed to get container status \"2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37\": rpc error: code = NotFound desc = could not find container \"2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37\": container with ID starting with 2e9295a417bba49a41a0e7ad55d19e2dbe6c2675665debec7974b4b1028ddc37 not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.433667 4945 scope.go:117] "RemoveContainer" containerID="534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.434184 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd\": container with ID starting with 534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd not found: ID does not exist" containerID="534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.434209 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd"} err="failed to get container status \"534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd\": rpc error: code = NotFound desc = could not find container \"534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd\": container with ID starting with 534e19d3cc201636623f29df587d02ff3d61565133ddfd645f2ae13cb602f2fd not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.434224 4945 scope.go:117] "RemoveContainer" containerID="a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.491062 4945 scope.go:117] "RemoveContainer" containerID="c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.505881 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-csg4k"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.509315 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-csg4k"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.524621 4945 scope.go:117] "RemoveContainer" containerID="cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.525674 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5w5w9"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.531437 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5w5w9"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.548495 4945 scope.go:117] "RemoveContainer" containerID="a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.549323 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0\": container with ID starting with a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0 not found: ID does not exist" containerID="a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.549373 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0"} err="failed to get container status \"a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0\": rpc error: code = NotFound desc = could not find container \"a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0\": container with ID starting with a0a262d7912202a96109da5d2c8c4b2937ac3725b2bd9581fa679108939adec0 not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.549411 4945 scope.go:117] "RemoveContainer" containerID="c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.550057 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a\": container with ID starting with c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a not found: ID does not exist" containerID="c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.550118 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a"} err="failed to get container status \"c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a\": rpc error: code = NotFound desc = could not find container \"c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a\": container with ID starting with c36ea368953ca7513548477be6b32bc354c51aa804e1682bda61e7884260eb1a not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.550162 4945 scope.go:117] "RemoveContainer" containerID="cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb" Dec 06 06:59:44 crc kubenswrapper[4945]: E1206 06:59:44.550623 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb\": container with ID starting with cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb not found: ID does not exist" containerID="cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.550663 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb"} err="failed to get container status \"cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb\": rpc error: code = NotFound desc = could not find container \"cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb\": container with ID starting with cef99eab4c475081367d597f12f16b4fd9540c0edac91d6a0a6d7c2752d157cb not found: ID does not exist" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.551510 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fvjzs"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.557089 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fvjzs"] Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.960414 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" path="/var/lib/kubelet/pods/319506ca-cc5d-424b-a84f-e7fab30c8a76/volumes" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.961481 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" path="/var/lib/kubelet/pods/4c50068a-69e4-4063-8f08-e4dc4e51fc8d/volumes" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.962164 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d7466cb-add7-436c-8641-0bf1bda1afbd" path="/var/lib/kubelet/pods/8d7466cb-add7-436c-8641-0bf1bda1afbd/volumes" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.963205 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" path="/var/lib/kubelet/pods/e315043e-3530-4af3-992b-8647e92bfa2d/volumes" Dec 06 06:59:44 crc kubenswrapper[4945]: I1206 06:59:44.963892 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" path="/var/lib/kubelet/pods/fb6b366b-20cd-404c-be4c-b07df65de3bb/volumes" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.196440 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" event={"ID":"5d136274-fa83-4411-845d-5d5cc3cff1b7","Type":"ContainerStarted","Data":"ea4e1f4eeaab8ce4a260aca3c7d5ec5277094f474c091ea792744a60dfda760a"} Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.196520 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" event={"ID":"5d136274-fa83-4411-845d-5d5cc3cff1b7","Type":"ContainerStarted","Data":"2fbe2b31b56b414d89341a39d4792f6507beeac5bf227dfbf0bc09ce5b1c3fe2"} Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.196928 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.200600 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.223974 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-6rgp5" podStartSLOduration=2.223950688 podStartE2EDuration="2.223950688s" podCreationTimestamp="2025-12-06 06:59:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:59:45.217191417 +0000 UTC m=+398.672052481" watchObservedRunningTime="2025-12-06 06:59:45.223950688 +0000 UTC m=+398.678811732" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457380 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gx58r"] Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457704 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457724 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457737 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7466cb-add7-436c-8641-0bf1bda1afbd" containerName="marketplace-operator" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457744 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7466cb-add7-436c-8641-0bf1bda1afbd" containerName="marketplace-operator" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457753 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerName="extract-utilities" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457760 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerName="extract-utilities" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457770 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457778 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457792 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" containerName="extract-utilities" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457798 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" containerName="extract-utilities" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457807 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerName="extract-utilities" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457813 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerName="extract-utilities" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457824 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" containerName="extract-content" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457831 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" containerName="extract-content" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457841 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerName="extract-utilities" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457848 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerName="extract-utilities" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457857 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457864 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457873 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457881 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457893 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerName="extract-content" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457900 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerName="extract-content" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457908 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerName="extract-content" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457915 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerName="extract-content" Dec 06 06:59:45 crc kubenswrapper[4945]: E1206 06:59:45.457927 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerName="extract-content" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.457938 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerName="extract-content" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.458061 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e315043e-3530-4af3-992b-8647e92bfa2d" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.458075 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7466cb-add7-436c-8641-0bf1bda1afbd" containerName="marketplace-operator" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.458089 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c50068a-69e4-4063-8f08-e4dc4e51fc8d" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.458102 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb6b366b-20cd-404c-be4c-b07df65de3bb" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.458112 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="319506ca-cc5d-424b-a84f-e7fab30c8a76" containerName="registry-server" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.459236 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.469585 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.474424 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gx58r"] Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.490632 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cmp4\" (UniqueName: \"kubernetes.io/projected/95ca28b7-42c4-4af4-8b14-d818ef8565e3-kube-api-access-9cmp4\") pod \"redhat-marketplace-gx58r\" (UID: \"95ca28b7-42c4-4af4-8b14-d818ef8565e3\") " pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.490714 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ca28b7-42c4-4af4-8b14-d818ef8565e3-catalog-content\") pod \"redhat-marketplace-gx58r\" (UID: \"95ca28b7-42c4-4af4-8b14-d818ef8565e3\") " pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.490741 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ca28b7-42c4-4af4-8b14-d818ef8565e3-utilities\") pod \"redhat-marketplace-gx58r\" (UID: \"95ca28b7-42c4-4af4-8b14-d818ef8565e3\") " pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.592326 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ca28b7-42c4-4af4-8b14-d818ef8565e3-utilities\") pod \"redhat-marketplace-gx58r\" (UID: \"95ca28b7-42c4-4af4-8b14-d818ef8565e3\") " pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.592482 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cmp4\" (UniqueName: \"kubernetes.io/projected/95ca28b7-42c4-4af4-8b14-d818ef8565e3-kube-api-access-9cmp4\") pod \"redhat-marketplace-gx58r\" (UID: \"95ca28b7-42c4-4af4-8b14-d818ef8565e3\") " pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.592537 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ca28b7-42c4-4af4-8b14-d818ef8565e3-catalog-content\") pod \"redhat-marketplace-gx58r\" (UID: \"95ca28b7-42c4-4af4-8b14-d818ef8565e3\") " pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.593625 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ca28b7-42c4-4af4-8b14-d818ef8565e3-utilities\") pod \"redhat-marketplace-gx58r\" (UID: \"95ca28b7-42c4-4af4-8b14-d818ef8565e3\") " pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.593703 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ca28b7-42c4-4af4-8b14-d818ef8565e3-catalog-content\") pod \"redhat-marketplace-gx58r\" (UID: \"95ca28b7-42c4-4af4-8b14-d818ef8565e3\") " pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.617493 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cmp4\" (UniqueName: \"kubernetes.io/projected/95ca28b7-42c4-4af4-8b14-d818ef8565e3-kube-api-access-9cmp4\") pod \"redhat-marketplace-gx58r\" (UID: \"95ca28b7-42c4-4af4-8b14-d818ef8565e3\") " pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.662163 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xgbc5"] Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.663325 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.671218 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.675374 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xgbc5"] Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.779609 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.795962 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a19579b9-98ab-40de-b39a-9419031a180c-catalog-content\") pod \"redhat-operators-xgbc5\" (UID: \"a19579b9-98ab-40de-b39a-9419031a180c\") " pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.796035 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrthc\" (UniqueName: \"kubernetes.io/projected/a19579b9-98ab-40de-b39a-9419031a180c-kube-api-access-mrthc\") pod \"redhat-operators-xgbc5\" (UID: \"a19579b9-98ab-40de-b39a-9419031a180c\") " pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.796085 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a19579b9-98ab-40de-b39a-9419031a180c-utilities\") pod \"redhat-operators-xgbc5\" (UID: \"a19579b9-98ab-40de-b39a-9419031a180c\") " pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.897497 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a19579b9-98ab-40de-b39a-9419031a180c-catalog-content\") pod \"redhat-operators-xgbc5\" (UID: \"a19579b9-98ab-40de-b39a-9419031a180c\") " pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.897556 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrthc\" (UniqueName: \"kubernetes.io/projected/a19579b9-98ab-40de-b39a-9419031a180c-kube-api-access-mrthc\") pod \"redhat-operators-xgbc5\" (UID: \"a19579b9-98ab-40de-b39a-9419031a180c\") " pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.897590 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a19579b9-98ab-40de-b39a-9419031a180c-utilities\") pod \"redhat-operators-xgbc5\" (UID: \"a19579b9-98ab-40de-b39a-9419031a180c\") " pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.898333 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a19579b9-98ab-40de-b39a-9419031a180c-utilities\") pod \"redhat-operators-xgbc5\" (UID: \"a19579b9-98ab-40de-b39a-9419031a180c\") " pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.898539 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a19579b9-98ab-40de-b39a-9419031a180c-catalog-content\") pod \"redhat-operators-xgbc5\" (UID: \"a19579b9-98ab-40de-b39a-9419031a180c\") " pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.922994 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrthc\" (UniqueName: \"kubernetes.io/projected/a19579b9-98ab-40de-b39a-9419031a180c-kube-api-access-mrthc\") pod \"redhat-operators-xgbc5\" (UID: \"a19579b9-98ab-40de-b39a-9419031a180c\") " pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:45 crc kubenswrapper[4945]: I1206 06:59:45.982534 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 06:59:46 crc kubenswrapper[4945]: I1206 06:59:46.228413 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gx58r"] Dec 06 06:59:46 crc kubenswrapper[4945]: W1206 06:59:46.236726 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95ca28b7_42c4_4af4_8b14_d818ef8565e3.slice/crio-e10d12d79d73446e0bec5124bc1b5a3bdb85f3ee41eb53e99ce0e5b1b90e1583 WatchSource:0}: Error finding container e10d12d79d73446e0bec5124bc1b5a3bdb85f3ee41eb53e99ce0e5b1b90e1583: Status 404 returned error can't find the container with id e10d12d79d73446e0bec5124bc1b5a3bdb85f3ee41eb53e99ce0e5b1b90e1583 Dec 06 06:59:46 crc kubenswrapper[4945]: I1206 06:59:46.409626 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xgbc5"] Dec 06 06:59:46 crc kubenswrapper[4945]: W1206 06:59:46.420652 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda19579b9_98ab_40de_b39a_9419031a180c.slice/crio-00fba5ea60b746909fcaaa29e7ad9aa9678468cf8f428cf38df34f8d9297303b WatchSource:0}: Error finding container 00fba5ea60b746909fcaaa29e7ad9aa9678468cf8f428cf38df34f8d9297303b: Status 404 returned error can't find the container with id 00fba5ea60b746909fcaaa29e7ad9aa9678468cf8f428cf38df34f8d9297303b Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.230727 4945 generic.go:334] "Generic (PLEG): container finished" podID="a19579b9-98ab-40de-b39a-9419031a180c" containerID="7c7f27ff027597feeb1f20ff96f2bf722fbdeecfcc1f168e5ab968d8f8de62b7" exitCode=0 Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.230832 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgbc5" event={"ID":"a19579b9-98ab-40de-b39a-9419031a180c","Type":"ContainerDied","Data":"7c7f27ff027597feeb1f20ff96f2bf722fbdeecfcc1f168e5ab968d8f8de62b7"} Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.231705 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgbc5" event={"ID":"a19579b9-98ab-40de-b39a-9419031a180c","Type":"ContainerStarted","Data":"00fba5ea60b746909fcaaa29e7ad9aa9678468cf8f428cf38df34f8d9297303b"} Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.233298 4945 generic.go:334] "Generic (PLEG): container finished" podID="95ca28b7-42c4-4af4-8b14-d818ef8565e3" containerID="a86b5c49a235ac5e2dae97a38db20b8be7f1cfc81be5743c90d70a3f382d0dfe" exitCode=0 Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.233424 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gx58r" event={"ID":"95ca28b7-42c4-4af4-8b14-d818ef8565e3","Type":"ContainerDied","Data":"a86b5c49a235ac5e2dae97a38db20b8be7f1cfc81be5743c90d70a3f382d0dfe"} Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.233458 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gx58r" event={"ID":"95ca28b7-42c4-4af4-8b14-d818ef8565e3","Type":"ContainerStarted","Data":"e10d12d79d73446e0bec5124bc1b5a3bdb85f3ee41eb53e99ce0e5b1b90e1583"} Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.860522 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rcv8f"] Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.862542 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.868443 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.874906 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rcv8f"] Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.933885 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8rv5\" (UniqueName: \"kubernetes.io/projected/417e92a7-1574-4c0a-ae51-c11449157595-kube-api-access-l8rv5\") pod \"certified-operators-rcv8f\" (UID: \"417e92a7-1574-4c0a-ae51-c11449157595\") " pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.933973 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/417e92a7-1574-4c0a-ae51-c11449157595-utilities\") pod \"certified-operators-rcv8f\" (UID: \"417e92a7-1574-4c0a-ae51-c11449157595\") " pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:47 crc kubenswrapper[4945]: I1206 06:59:47.934589 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/417e92a7-1574-4c0a-ae51-c11449157595-catalog-content\") pod \"certified-operators-rcv8f\" (UID: \"417e92a7-1574-4c0a-ae51-c11449157595\") " pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.036517 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/417e92a7-1574-4c0a-ae51-c11449157595-catalog-content\") pod \"certified-operators-rcv8f\" (UID: \"417e92a7-1574-4c0a-ae51-c11449157595\") " pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.036707 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8rv5\" (UniqueName: \"kubernetes.io/projected/417e92a7-1574-4c0a-ae51-c11449157595-kube-api-access-l8rv5\") pod \"certified-operators-rcv8f\" (UID: \"417e92a7-1574-4c0a-ae51-c11449157595\") " pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.036748 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/417e92a7-1574-4c0a-ae51-c11449157595-utilities\") pod \"certified-operators-rcv8f\" (UID: \"417e92a7-1574-4c0a-ae51-c11449157595\") " pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.037384 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/417e92a7-1574-4c0a-ae51-c11449157595-catalog-content\") pod \"certified-operators-rcv8f\" (UID: \"417e92a7-1574-4c0a-ae51-c11449157595\") " pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.037502 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/417e92a7-1574-4c0a-ae51-c11449157595-utilities\") pod \"certified-operators-rcv8f\" (UID: \"417e92a7-1574-4c0a-ae51-c11449157595\") " pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.064961 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8rv5\" (UniqueName: \"kubernetes.io/projected/417e92a7-1574-4c0a-ae51-c11449157595-kube-api-access-l8rv5\") pod \"certified-operators-rcv8f\" (UID: \"417e92a7-1574-4c0a-ae51-c11449157595\") " pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.066323 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n9zht"] Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.067502 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.071163 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.079256 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n9zht"] Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.137978 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj75j\" (UniqueName: \"kubernetes.io/projected/17d087a6-7ecb-4b21-b825-49fd80f6faf4-kube-api-access-kj75j\") pod \"community-operators-n9zht\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.138086 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-catalog-content\") pod \"community-operators-n9zht\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.138146 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-utilities\") pod \"community-operators-n9zht\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.203925 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.239821 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-catalog-content\") pod \"community-operators-n9zht\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.239913 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-utilities\") pod \"community-operators-n9zht\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.239992 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj75j\" (UniqueName: \"kubernetes.io/projected/17d087a6-7ecb-4b21-b825-49fd80f6faf4-kube-api-access-kj75j\") pod \"community-operators-n9zht\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.241108 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-catalog-content\") pod \"community-operators-n9zht\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.241595 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgbc5" event={"ID":"a19579b9-98ab-40de-b39a-9419031a180c","Type":"ContainerStarted","Data":"58a37dda5eec2be12e8a781f237dfbfc1a6510f628bb2b0e6a8fa1c85332f197"} Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.241656 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-utilities\") pod \"community-operators-n9zht\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.248074 4945 generic.go:334] "Generic (PLEG): container finished" podID="95ca28b7-42c4-4af4-8b14-d818ef8565e3" containerID="3f4e765102688de9d5db98bbba2bd735a76c15703a4aeeedbdc482576b2ca039" exitCode=0 Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.248156 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gx58r" event={"ID":"95ca28b7-42c4-4af4-8b14-d818ef8565e3","Type":"ContainerDied","Data":"3f4e765102688de9d5db98bbba2bd735a76c15703a4aeeedbdc482576b2ca039"} Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.273383 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj75j\" (UniqueName: \"kubernetes.io/projected/17d087a6-7ecb-4b21-b825-49fd80f6faf4-kube-api-access-kj75j\") pod \"community-operators-n9zht\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.408875 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9zht" Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.678829 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rcv8f"] Dec 06 06:59:48 crc kubenswrapper[4945]: I1206 06:59:48.834155 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n9zht"] Dec 06 06:59:48 crc kubenswrapper[4945]: W1206 06:59:48.838555 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17d087a6_7ecb_4b21_b825_49fd80f6faf4.slice/crio-8a30ec4ffdad40b4a9860ed83d4b3f59d37d72946ded1ad633d1193c0f18868d WatchSource:0}: Error finding container 8a30ec4ffdad40b4a9860ed83d4b3f59d37d72946ded1ad633d1193c0f18868d: Status 404 returned error can't find the container with id 8a30ec4ffdad40b4a9860ed83d4b3f59d37d72946ded1ad633d1193c0f18868d Dec 06 06:59:49 crc kubenswrapper[4945]: I1206 06:59:49.257122 4945 generic.go:334] "Generic (PLEG): container finished" podID="a19579b9-98ab-40de-b39a-9419031a180c" containerID="58a37dda5eec2be12e8a781f237dfbfc1a6510f628bb2b0e6a8fa1c85332f197" exitCode=0 Dec 06 06:59:49 crc kubenswrapper[4945]: I1206 06:59:49.257230 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgbc5" event={"ID":"a19579b9-98ab-40de-b39a-9419031a180c","Type":"ContainerDied","Data":"58a37dda5eec2be12e8a781f237dfbfc1a6510f628bb2b0e6a8fa1c85332f197"} Dec 06 06:59:49 crc kubenswrapper[4945]: I1206 06:59:49.259506 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9zht" event={"ID":"17d087a6-7ecb-4b21-b825-49fd80f6faf4","Type":"ContainerStarted","Data":"8a30ec4ffdad40b4a9860ed83d4b3f59d37d72946ded1ad633d1193c0f18868d"} Dec 06 06:59:49 crc kubenswrapper[4945]: I1206 06:59:49.272940 4945 generic.go:334] "Generic (PLEG): container finished" podID="417e92a7-1574-4c0a-ae51-c11449157595" containerID="7b2f7c720668e1b4c6beffb84a65f097bb0d140317a2008d4129508262196395" exitCode=0 Dec 06 06:59:49 crc kubenswrapper[4945]: I1206 06:59:49.273037 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcv8f" event={"ID":"417e92a7-1574-4c0a-ae51-c11449157595","Type":"ContainerDied","Data":"7b2f7c720668e1b4c6beffb84a65f097bb0d140317a2008d4129508262196395"} Dec 06 06:59:49 crc kubenswrapper[4945]: I1206 06:59:49.273080 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcv8f" event={"ID":"417e92a7-1574-4c0a-ae51-c11449157595","Type":"ContainerStarted","Data":"faa3dcae5eab5f657dbcdb7dddb39b7beaa408a0656b6a99eb204d08b5ece8d6"} Dec 06 06:59:52 crc kubenswrapper[4945]: I1206 06:59:52.303419 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gx58r" event={"ID":"95ca28b7-42c4-4af4-8b14-d818ef8565e3","Type":"ContainerStarted","Data":"94200dd08c1e1a59b12e7783130aa5a1be7226045ae8232a0316b0782fc8d323"} Dec 06 06:59:52 crc kubenswrapper[4945]: I1206 06:59:52.307566 4945 generic.go:334] "Generic (PLEG): container finished" podID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerID="bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9" exitCode=0 Dec 06 06:59:52 crc kubenswrapper[4945]: I1206 06:59:52.307669 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9zht" event={"ID":"17d087a6-7ecb-4b21-b825-49fd80f6faf4","Type":"ContainerDied","Data":"bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9"} Dec 06 06:59:52 crc kubenswrapper[4945]: I1206 06:59:52.351036 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gx58r" podStartSLOduration=5.675828409 podStartE2EDuration="7.351006686s" podCreationTimestamp="2025-12-06 06:59:45 +0000 UTC" firstStartedPulling="2025-12-06 06:59:47.235766066 +0000 UTC m=+400.690627110" lastFinishedPulling="2025-12-06 06:59:48.910944343 +0000 UTC m=+402.365805387" observedRunningTime="2025-12-06 06:59:52.343504905 +0000 UTC m=+405.798365979" watchObservedRunningTime="2025-12-06 06:59:52.351006686 +0000 UTC m=+405.805867740" Dec 06 06:59:54 crc kubenswrapper[4945]: I1206 06:59:54.325201 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgbc5" event={"ID":"a19579b9-98ab-40de-b39a-9419031a180c","Type":"ContainerStarted","Data":"631d4d2f54515437d6d090d48b509c09d4c017744ef2f42d28fa92093389685e"} Dec 06 06:59:54 crc kubenswrapper[4945]: I1206 06:59:54.330245 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-lslh4" Dec 06 06:59:54 crc kubenswrapper[4945]: I1206 06:59:54.390835 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8ppq"] Dec 06 06:59:55 crc kubenswrapper[4945]: I1206 06:59:55.332750 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcv8f" event={"ID":"417e92a7-1574-4c0a-ae51-c11449157595","Type":"ContainerStarted","Data":"73ff3c15a2c31b91bcf9ec2b87b9295e7ea75ee1fae457625cd5ce3f82448f0c"} Dec 06 06:59:55 crc kubenswrapper[4945]: I1206 06:59:55.780333 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:55 crc kubenswrapper[4945]: I1206 06:59:55.781406 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:55 crc kubenswrapper[4945]: I1206 06:59:55.822610 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 06:59:56 crc kubenswrapper[4945]: I1206 06:59:56.340881 4945 generic.go:334] "Generic (PLEG): container finished" podID="417e92a7-1574-4c0a-ae51-c11449157595" containerID="73ff3c15a2c31b91bcf9ec2b87b9295e7ea75ee1fae457625cd5ce3f82448f0c" exitCode=0 Dec 06 06:59:56 crc kubenswrapper[4945]: I1206 06:59:56.340985 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcv8f" event={"ID":"417e92a7-1574-4c0a-ae51-c11449157595","Type":"ContainerDied","Data":"73ff3c15a2c31b91bcf9ec2b87b9295e7ea75ee1fae457625cd5ce3f82448f0c"} Dec 06 06:59:56 crc kubenswrapper[4945]: I1206 06:59:56.385548 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xgbc5" podStartSLOduration=6.01697252 podStartE2EDuration="11.385516592s" podCreationTimestamp="2025-12-06 06:59:45 +0000 UTC" firstStartedPulling="2025-12-06 06:59:47.234827821 +0000 UTC m=+400.689688865" lastFinishedPulling="2025-12-06 06:59:52.603371893 +0000 UTC m=+406.058232937" observedRunningTime="2025-12-06 06:59:56.384019102 +0000 UTC m=+409.838880156" watchObservedRunningTime="2025-12-06 06:59:56.385516592 +0000 UTC m=+409.840377646" Dec 06 06:59:56 crc kubenswrapper[4945]: I1206 06:59:56.398325 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gx58r" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.181429 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw"] Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.183160 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.187038 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.187242 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.196565 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw"] Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.219421 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-448ds\" (UniqueName: \"kubernetes.io/projected/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-kube-api-access-448ds\") pod \"collect-profiles-29416740-mfhxw\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.219504 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-config-volume\") pod \"collect-profiles-29416740-mfhxw\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.219640 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-secret-volume\") pod \"collect-profiles-29416740-mfhxw\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.320760 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-secret-volume\") pod \"collect-profiles-29416740-mfhxw\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.320829 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-448ds\" (UniqueName: \"kubernetes.io/projected/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-kube-api-access-448ds\") pod \"collect-profiles-29416740-mfhxw\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.320859 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-config-volume\") pod \"collect-profiles-29416740-mfhxw\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.322037 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-config-volume\") pod \"collect-profiles-29416740-mfhxw\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.329086 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-secret-volume\") pod \"collect-profiles-29416740-mfhxw\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.336562 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-448ds\" (UniqueName: \"kubernetes.io/projected/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-kube-api-access-448ds\") pod \"collect-profiles-29416740-mfhxw\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.504240 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:00 crc kubenswrapper[4945]: I1206 07:00:00.919893 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw"] Dec 06 07:00:01 crc kubenswrapper[4945]: I1206 07:00:01.383098 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" event={"ID":"7c7767b5-806a-47a0-8ec6-be3e9fdeca18","Type":"ContainerStarted","Data":"529adbd03fce3223ae33f9058b3afbecc27fb7317394306ae137f5772e02fe5a"} Dec 06 07:00:05 crc kubenswrapper[4945]: I1206 07:00:05.405675 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" event={"ID":"7c7767b5-806a-47a0-8ec6-be3e9fdeca18","Type":"ContainerStarted","Data":"e13824d49ddbecada56e31c77cd25010844935e9d016a2d8e0ba5c61c06f381d"} Dec 06 07:00:05 crc kubenswrapper[4945]: I1206 07:00:05.983531 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 07:00:05 crc kubenswrapper[4945]: I1206 07:00:05.983760 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 07:00:06 crc kubenswrapper[4945]: I1206 07:00:06.043955 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 07:00:06 crc kubenswrapper[4945]: I1206 07:00:06.431014 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" podStartSLOduration=6.430994697 podStartE2EDuration="6.430994697s" podCreationTimestamp="2025-12-06 07:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:00:06.427714079 +0000 UTC m=+419.882575123" watchObservedRunningTime="2025-12-06 07:00:06.430994697 +0000 UTC m=+419.885855741" Dec 06 07:00:06 crc kubenswrapper[4945]: I1206 07:00:06.457179 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xgbc5" Dec 06 07:00:07 crc kubenswrapper[4945]: I1206 07:00:07.037762 4945 scope.go:117] "RemoveContainer" containerID="90ef3334330ca39386c527ce30ac7d0868280d607d2da31a8c8eab826b46b6a5" Dec 06 07:00:07 crc kubenswrapper[4945]: I1206 07:00:07.799577 4945 scope.go:117] "RemoveContainer" containerID="83b6bd76c21613b8c3fc008275ebe0b6c8adf39a5fa1167fcfe968b3fd4600f2" Dec 06 07:00:07 crc kubenswrapper[4945]: I1206 07:00:07.930895 4945 scope.go:117] "RemoveContainer" containerID="5dbdbd40467d55b66b81460f4ffb3b1643f19d9667961d88ad4eda6161ad88ac" Dec 06 07:00:07 crc kubenswrapper[4945]: I1206 07:00:07.966863 4945 scope.go:117] "RemoveContainer" containerID="b8dbcd249349300f698ea11b10c32bfe84b707860e18845026f349a43bed5e41" Dec 06 07:00:07 crc kubenswrapper[4945]: I1206 07:00:07.996690 4945 scope.go:117] "RemoveContainer" containerID="d58814ab2be8b6d79537ebde38315e1c82e1f56d5785919158bb1019217872cb" Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.026396 4945 scope.go:117] "RemoveContainer" containerID="2b5beacd070dc86234d8673185a54d650ad70146e896a19cd423825fb41f1fa5" Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.424385 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9zht" event={"ID":"17d087a6-7ecb-4b21-b825-49fd80f6faf4","Type":"ContainerStarted","Data":"ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea"} Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.426317 4945 generic.go:334] "Generic (PLEG): container finished" podID="7c7767b5-806a-47a0-8ec6-be3e9fdeca18" containerID="e13824d49ddbecada56e31c77cd25010844935e9d016a2d8e0ba5c61c06f381d" exitCode=0 Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.426389 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" event={"ID":"7c7767b5-806a-47a0-8ec6-be3e9fdeca18","Type":"ContainerDied","Data":"e13824d49ddbecada56e31c77cd25010844935e9d016a2d8e0ba5c61c06f381d"} Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.428674 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcv8f" event={"ID":"417e92a7-1574-4c0a-ae51-c11449157595","Type":"ContainerStarted","Data":"2f73436973fd3a824e19ac5d27e324419a17795d85c38e08e1af7b3cb07121c6"} Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.477857 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rcv8f" podStartSLOduration=2.8173288359999997 podStartE2EDuration="21.477838432s" podCreationTimestamp="2025-12-06 06:59:47 +0000 UTC" firstStartedPulling="2025-12-06 06:59:49.277527024 +0000 UTC m=+402.732388068" lastFinishedPulling="2025-12-06 07:00:07.93803662 +0000 UTC m=+421.392897664" observedRunningTime="2025-12-06 07:00:08.47553487 +0000 UTC m=+421.930395924" watchObservedRunningTime="2025-12-06 07:00:08.477838432 +0000 UTC m=+421.932699476" Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.796108 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.796222 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.796343 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.797665 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c0055ca88a5066f06888e8872d8e2fb3e70f24ff64dd4e6919f6a317fcc06177"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:00:08 crc kubenswrapper[4945]: I1206 07:00:08.797811 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://c0055ca88a5066f06888e8872d8e2fb3e70f24ff64dd4e6919f6a317fcc06177" gracePeriod=600 Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.436122 4945 generic.go:334] "Generic (PLEG): container finished" podID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerID="ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea" exitCode=0 Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.436203 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9zht" event={"ID":"17d087a6-7ecb-4b21-b825-49fd80f6faf4","Type":"ContainerDied","Data":"ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea"} Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.438906 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="c0055ca88a5066f06888e8872d8e2fb3e70f24ff64dd4e6919f6a317fcc06177" exitCode=0 Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.438971 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"c0055ca88a5066f06888e8872d8e2fb3e70f24ff64dd4e6919f6a317fcc06177"} Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.439012 4945 scope.go:117] "RemoveContainer" containerID="0349b3e5753fd1eeabafe6bab6ad1df2063a7d3144cc6db520d7a4688f967b02" Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.777905 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.864077 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-secret-volume\") pod \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.864137 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-config-volume\") pod \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.864195 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-448ds\" (UniqueName: \"kubernetes.io/projected/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-kube-api-access-448ds\") pod \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\" (UID: \"7c7767b5-806a-47a0-8ec6-be3e9fdeca18\") " Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.865255 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-config-volume" (OuterVolumeSpecName: "config-volume") pod "7c7767b5-806a-47a0-8ec6-be3e9fdeca18" (UID: "7c7767b5-806a-47a0-8ec6-be3e9fdeca18"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.871950 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7c7767b5-806a-47a0-8ec6-be3e9fdeca18" (UID: "7c7767b5-806a-47a0-8ec6-be3e9fdeca18"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.872009 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-kube-api-access-448ds" (OuterVolumeSpecName: "kube-api-access-448ds") pod "7c7767b5-806a-47a0-8ec6-be3e9fdeca18" (UID: "7c7767b5-806a-47a0-8ec6-be3e9fdeca18"). InnerVolumeSpecName "kube-api-access-448ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.966014 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.966057 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:09 crc kubenswrapper[4945]: I1206 07:00:09.966068 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-448ds\" (UniqueName: \"kubernetes.io/projected/7c7767b5-806a-47a0-8ec6-be3e9fdeca18-kube-api-access-448ds\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:10 crc kubenswrapper[4945]: I1206 07:00:10.452359 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"845576fa1982d7e18bba49498a66f9811779a237678acfa18ccbe96601cf2b7b"} Dec 06 07:00:10 crc kubenswrapper[4945]: I1206 07:00:10.456801 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" event={"ID":"7c7767b5-806a-47a0-8ec6-be3e9fdeca18","Type":"ContainerDied","Data":"529adbd03fce3223ae33f9058b3afbecc27fb7317394306ae137f5772e02fe5a"} Dec 06 07:00:10 crc kubenswrapper[4945]: I1206 07:00:10.456836 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="529adbd03fce3223ae33f9058b3afbecc27fb7317394306ae137f5772e02fe5a" Dec 06 07:00:10 crc kubenswrapper[4945]: I1206 07:00:10.456902 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw" Dec 06 07:00:11 crc kubenswrapper[4945]: I1206 07:00:11.466802 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9zht" event={"ID":"17d087a6-7ecb-4b21-b825-49fd80f6faf4","Type":"ContainerStarted","Data":"163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186"} Dec 06 07:00:11 crc kubenswrapper[4945]: I1206 07:00:11.490466 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n9zht" podStartSLOduration=5.427455196 podStartE2EDuration="23.490442526s" podCreationTimestamp="2025-12-06 06:59:48 +0000 UTC" firstStartedPulling="2025-12-06 06:59:52.31188092 +0000 UTC m=+405.766741964" lastFinishedPulling="2025-12-06 07:00:10.37486825 +0000 UTC m=+423.829729294" observedRunningTime="2025-12-06 07:00:11.489752828 +0000 UTC m=+424.944613872" watchObservedRunningTime="2025-12-06 07:00:11.490442526 +0000 UTC m=+424.945303560" Dec 06 07:00:18 crc kubenswrapper[4945]: I1206 07:00:18.205077 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 07:00:18 crc kubenswrapper[4945]: I1206 07:00:18.205769 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 07:00:18 crc kubenswrapper[4945]: I1206 07:00:18.253434 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 07:00:18 crc kubenswrapper[4945]: I1206 07:00:18.409629 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n9zht" Dec 06 07:00:18 crc kubenswrapper[4945]: I1206 07:00:18.409748 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n9zht" Dec 06 07:00:18 crc kubenswrapper[4945]: I1206 07:00:18.450233 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n9zht" Dec 06 07:00:18 crc kubenswrapper[4945]: I1206 07:00:18.545202 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n9zht" Dec 06 07:00:18 crc kubenswrapper[4945]: I1206 07:00:18.551053 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rcv8f" Dec 06 07:00:19 crc kubenswrapper[4945]: I1206 07:00:19.439420 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" podUID="401f27c7-5383-4ade-839b-59a92b137752" containerName="registry" containerID="cri-o://a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4" gracePeriod=30 Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.014191 4945 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-k8ppq container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.35:5000/healthz\": dial tcp 10.217.0.35:5000: connect: connection refused" start-of-body= Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.014867 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" podUID="401f27c7-5383-4ade-839b-59a92b137752" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.35:5000/healthz\": dial tcp 10.217.0.35:5000: connect: connection refused" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.471459 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.513821 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-trusted-ca\") pod \"401f27c7-5383-4ade-839b-59a92b137752\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.513879 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-registry-certificates\") pod \"401f27c7-5383-4ade-839b-59a92b137752\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.513937 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-bound-sa-token\") pod \"401f27c7-5383-4ade-839b-59a92b137752\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.514016 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/401f27c7-5383-4ade-839b-59a92b137752-ca-trust-extracted\") pod \"401f27c7-5383-4ade-839b-59a92b137752\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.514044 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/401f27c7-5383-4ade-839b-59a92b137752-installation-pull-secrets\") pod \"401f27c7-5383-4ade-839b-59a92b137752\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.514167 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"401f27c7-5383-4ade-839b-59a92b137752\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.514194 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcmnq\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-kube-api-access-pcmnq\") pod \"401f27c7-5383-4ade-839b-59a92b137752\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.514236 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-registry-tls\") pod \"401f27c7-5383-4ade-839b-59a92b137752\" (UID: \"401f27c7-5383-4ade-839b-59a92b137752\") " Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.515081 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "401f27c7-5383-4ade-839b-59a92b137752" (UID: "401f27c7-5383-4ade-839b-59a92b137752"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.515100 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "401f27c7-5383-4ade-839b-59a92b137752" (UID: "401f27c7-5383-4ade-839b-59a92b137752"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.515561 4945 generic.go:334] "Generic (PLEG): container finished" podID="401f27c7-5383-4ade-839b-59a92b137752" containerID="a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4" exitCode=0 Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.515599 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" event={"ID":"401f27c7-5383-4ade-839b-59a92b137752","Type":"ContainerDied","Data":"a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4"} Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.515624 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" event={"ID":"401f27c7-5383-4ade-839b-59a92b137752","Type":"ContainerDied","Data":"1a4f727d0af718df8e80387b21d10b276779c17a4e78ccaa887430fe783ede59"} Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.515641 4945 scope.go:117] "RemoveContainer" containerID="a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.515749 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k8ppq" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.521112 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401f27c7-5383-4ade-839b-59a92b137752-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "401f27c7-5383-4ade-839b-59a92b137752" (UID: "401f27c7-5383-4ade-839b-59a92b137752"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.522102 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "401f27c7-5383-4ade-839b-59a92b137752" (UID: "401f27c7-5383-4ade-839b-59a92b137752"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.523129 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-kube-api-access-pcmnq" (OuterVolumeSpecName: "kube-api-access-pcmnq") pod "401f27c7-5383-4ade-839b-59a92b137752" (UID: "401f27c7-5383-4ade-839b-59a92b137752"). InnerVolumeSpecName "kube-api-access-pcmnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.524710 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "401f27c7-5383-4ade-839b-59a92b137752" (UID: "401f27c7-5383-4ade-839b-59a92b137752"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.526134 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "401f27c7-5383-4ade-839b-59a92b137752" (UID: "401f27c7-5383-4ade-839b-59a92b137752"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.539401 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/401f27c7-5383-4ade-839b-59a92b137752-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "401f27c7-5383-4ade-839b-59a92b137752" (UID: "401f27c7-5383-4ade-839b-59a92b137752"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.557293 4945 scope.go:117] "RemoveContainer" containerID="a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4" Dec 06 07:00:20 crc kubenswrapper[4945]: E1206 07:00:20.557901 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4\": container with ID starting with a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4 not found: ID does not exist" containerID="a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.557968 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4"} err="failed to get container status \"a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4\": rpc error: code = NotFound desc = could not find container \"a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4\": container with ID starting with a96e3890285931dd1aa5d7e3ac4291f4e8fc71c17becb6f448831b86043dd3a4 not found: ID does not exist" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.615752 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcmnq\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-kube-api-access-pcmnq\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.615808 4945 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.615826 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.615836 4945 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/401f27c7-5383-4ade-839b-59a92b137752-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.615847 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/401f27c7-5383-4ade-839b-59a92b137752-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.615863 4945 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/401f27c7-5383-4ade-839b-59a92b137752-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.615876 4945 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/401f27c7-5383-4ade-839b-59a92b137752-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.843318 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8ppq"] Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.848140 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8ppq"] Dec 06 07:00:20 crc kubenswrapper[4945]: I1206 07:00:20.960268 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="401f27c7-5383-4ade-839b-59a92b137752" path="/var/lib/kubelet/pods/401f27c7-5383-4ade-839b-59a92b137752/volumes" Dec 06 07:02:38 crc kubenswrapper[4945]: I1206 07:02:38.795829 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:02:38 crc kubenswrapper[4945]: I1206 07:02:38.796362 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:03:08 crc kubenswrapper[4945]: I1206 07:03:08.796275 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:03:08 crc kubenswrapper[4945]: I1206 07:03:08.797087 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:03:38 crc kubenswrapper[4945]: I1206 07:03:38.795530 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:03:38 crc kubenswrapper[4945]: I1206 07:03:38.796139 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:03:38 crc kubenswrapper[4945]: I1206 07:03:38.796196 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:03:38 crc kubenswrapper[4945]: I1206 07:03:38.796754 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"845576fa1982d7e18bba49498a66f9811779a237678acfa18ccbe96601cf2b7b"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:03:38 crc kubenswrapper[4945]: I1206 07:03:38.796810 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://845576fa1982d7e18bba49498a66f9811779a237678acfa18ccbe96601cf2b7b" gracePeriod=600 Dec 06 07:03:39 crc kubenswrapper[4945]: I1206 07:03:39.881753 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="845576fa1982d7e18bba49498a66f9811779a237678acfa18ccbe96601cf2b7b" exitCode=0 Dec 06 07:03:39 crc kubenswrapper[4945]: I1206 07:03:39.881832 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"845576fa1982d7e18bba49498a66f9811779a237678acfa18ccbe96601cf2b7b"} Dec 06 07:03:39 crc kubenswrapper[4945]: I1206 07:03:39.882084 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"0c1b134b46f37e7ead92dfbee219cc6a32bd1e1e7b37766e6910364e7c42deb2"} Dec 06 07:03:39 crc kubenswrapper[4945]: I1206 07:03:39.882106 4945 scope.go:117] "RemoveContainer" containerID="c0055ca88a5066f06888e8872d8e2fb3e70f24ff64dd4e6919f6a317fcc06177" Dec 06 07:05:42 crc kubenswrapper[4945]: I1206 07:05:42.735403 4945 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 07:06:08 crc kubenswrapper[4945]: I1206 07:06:08.795755 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:06:08 crc kubenswrapper[4945]: I1206 07:06:08.796403 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:06:38 crc kubenswrapper[4945]: I1206 07:06:38.795647 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:06:38 crc kubenswrapper[4945]: I1206 07:06:38.796686 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.350803 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zqfs9"] Dec 06 07:06:55 crc kubenswrapper[4945]: E1206 07:06:55.351361 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401f27c7-5383-4ade-839b-59a92b137752" containerName="registry" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.351376 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="401f27c7-5383-4ade-839b-59a92b137752" containerName="registry" Dec 06 07:06:55 crc kubenswrapper[4945]: E1206 07:06:55.351398 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c7767b5-806a-47a0-8ec6-be3e9fdeca18" containerName="collect-profiles" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.351404 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c7767b5-806a-47a0-8ec6-be3e9fdeca18" containerName="collect-profiles" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.351501 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c7767b5-806a-47a0-8ec6-be3e9fdeca18" containerName="collect-profiles" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.351515 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="401f27c7-5383-4ade-839b-59a92b137752" containerName="registry" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.352429 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.365772 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zqfs9"] Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.469629 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-utilities\") pod \"certified-operators-zqfs9\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.469923 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-catalog-content\") pod \"certified-operators-zqfs9\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.469968 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2nhf\" (UniqueName: \"kubernetes.io/projected/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-kube-api-access-t2nhf\") pod \"certified-operators-zqfs9\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.571049 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-catalog-content\") pod \"certified-operators-zqfs9\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.571118 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2nhf\" (UniqueName: \"kubernetes.io/projected/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-kube-api-access-t2nhf\") pod \"certified-operators-zqfs9\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.571144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-utilities\") pod \"certified-operators-zqfs9\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.571746 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-utilities\") pod \"certified-operators-zqfs9\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.572030 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-catalog-content\") pod \"certified-operators-zqfs9\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.592616 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2nhf\" (UniqueName: \"kubernetes.io/projected/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-kube-api-access-t2nhf\") pod \"certified-operators-zqfs9\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.672452 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:06:55 crc kubenswrapper[4945]: I1206 07:06:55.933112 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zqfs9"] Dec 06 07:06:56 crc kubenswrapper[4945]: I1206 07:06:56.031564 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zqfs9" event={"ID":"2dbee646-d4c4-4ae1-baf1-36736ff75c9e","Type":"ContainerStarted","Data":"a79ad9c7e8e97261b1f64f478273d26c41fb78b1af58662207622c50ba33cb34"} Dec 06 07:06:57 crc kubenswrapper[4945]: I1206 07:06:57.037895 4945 generic.go:334] "Generic (PLEG): container finished" podID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerID="74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480" exitCode=0 Dec 06 07:06:57 crc kubenswrapper[4945]: I1206 07:06:57.038006 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zqfs9" event={"ID":"2dbee646-d4c4-4ae1-baf1-36736ff75c9e","Type":"ContainerDied","Data":"74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480"} Dec 06 07:06:57 crc kubenswrapper[4945]: I1206 07:06:57.040979 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:06:58 crc kubenswrapper[4945]: I1206 07:06:58.047325 4945 generic.go:334] "Generic (PLEG): container finished" podID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerID="eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde" exitCode=0 Dec 06 07:06:58 crc kubenswrapper[4945]: I1206 07:06:58.047465 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zqfs9" event={"ID":"2dbee646-d4c4-4ae1-baf1-36736ff75c9e","Type":"ContainerDied","Data":"eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde"} Dec 06 07:06:59 crc kubenswrapper[4945]: I1206 07:06:59.056948 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zqfs9" event={"ID":"2dbee646-d4c4-4ae1-baf1-36736ff75c9e","Type":"ContainerStarted","Data":"85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2"} Dec 06 07:06:59 crc kubenswrapper[4945]: I1206 07:06:59.084914 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zqfs9" podStartSLOduration=2.656179913 podStartE2EDuration="4.084885657s" podCreationTimestamp="2025-12-06 07:06:55 +0000 UTC" firstStartedPulling="2025-12-06 07:06:57.040693466 +0000 UTC m=+830.495554510" lastFinishedPulling="2025-12-06 07:06:58.46939921 +0000 UTC m=+831.924260254" observedRunningTime="2025-12-06 07:06:59.079932473 +0000 UTC m=+832.534793537" watchObservedRunningTime="2025-12-06 07:06:59.084885657 +0000 UTC m=+832.539746701" Dec 06 07:07:05 crc kubenswrapper[4945]: I1206 07:07:05.673139 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:07:05 crc kubenswrapper[4945]: I1206 07:07:05.673955 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:07:05 crc kubenswrapper[4945]: I1206 07:07:05.723702 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:07:06 crc kubenswrapper[4945]: I1206 07:07:06.147544 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:07:06 crc kubenswrapper[4945]: I1206 07:07:06.199033 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zqfs9"] Dec 06 07:07:08 crc kubenswrapper[4945]: I1206 07:07:08.118057 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zqfs9" podUID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerName="registry-server" containerID="cri-o://85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2" gracePeriod=2 Dec 06 07:07:08 crc kubenswrapper[4945]: I1206 07:07:08.796008 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:07:08 crc kubenswrapper[4945]: I1206 07:07:08.796684 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:07:08 crc kubenswrapper[4945]: I1206 07:07:08.796771 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:07:08 crc kubenswrapper[4945]: I1206 07:07:08.797917 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c1b134b46f37e7ead92dfbee219cc6a32bd1e1e7b37766e6910364e7c42deb2"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:07:08 crc kubenswrapper[4945]: I1206 07:07:08.798009 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://0c1b134b46f37e7ead92dfbee219cc6a32bd1e1e7b37766e6910364e7c42deb2" gracePeriod=600 Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.008650 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.127209 4945 generic.go:334] "Generic (PLEG): container finished" podID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerID="85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2" exitCode=0 Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.127456 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zqfs9" event={"ID":"2dbee646-d4c4-4ae1-baf1-36736ff75c9e","Type":"ContainerDied","Data":"85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2"} Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.127886 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zqfs9" event={"ID":"2dbee646-d4c4-4ae1-baf1-36736ff75c9e","Type":"ContainerDied","Data":"a79ad9c7e8e97261b1f64f478273d26c41fb78b1af58662207622c50ba33cb34"} Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.127591 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zqfs9" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.127955 4945 scope.go:117] "RemoveContainer" containerID="85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.133036 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="0c1b134b46f37e7ead92dfbee219cc6a32bd1e1e7b37766e6910364e7c42deb2" exitCode=0 Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.133093 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"0c1b134b46f37e7ead92dfbee219cc6a32bd1e1e7b37766e6910364e7c42deb2"} Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.133132 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"fcc05253d8d77edb32c43d3ca8c759b1e83d7e91658c6a423615205b104949ce"} Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.146589 4945 scope.go:117] "RemoveContainer" containerID="eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.159840 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-utilities\") pod \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.159907 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2nhf\" (UniqueName: \"kubernetes.io/projected/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-kube-api-access-t2nhf\") pod \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.159997 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-catalog-content\") pod \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\" (UID: \"2dbee646-d4c4-4ae1-baf1-36736ff75c9e\") " Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.160974 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-utilities" (OuterVolumeSpecName: "utilities") pod "2dbee646-d4c4-4ae1-baf1-36736ff75c9e" (UID: "2dbee646-d4c4-4ae1-baf1-36736ff75c9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.167757 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-kube-api-access-t2nhf" (OuterVolumeSpecName: "kube-api-access-t2nhf") pod "2dbee646-d4c4-4ae1-baf1-36736ff75c9e" (UID: "2dbee646-d4c4-4ae1-baf1-36736ff75c9e"). InnerVolumeSpecName "kube-api-access-t2nhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.171449 4945 scope.go:117] "RemoveContainer" containerID="74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.189353 4945 scope.go:117] "RemoveContainer" containerID="85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2" Dec 06 07:07:09 crc kubenswrapper[4945]: E1206 07:07:09.189952 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2\": container with ID starting with 85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2 not found: ID does not exist" containerID="85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.190006 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2"} err="failed to get container status \"85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2\": rpc error: code = NotFound desc = could not find container \"85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2\": container with ID starting with 85862fc98bfa286e7a3563bab16e245240c0a113f4266852e2a99960a67f47e2 not found: ID does not exist" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.190042 4945 scope.go:117] "RemoveContainer" containerID="eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde" Dec 06 07:07:09 crc kubenswrapper[4945]: E1206 07:07:09.190467 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde\": container with ID starting with eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde not found: ID does not exist" containerID="eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.190494 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde"} err="failed to get container status \"eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde\": rpc error: code = NotFound desc = could not find container \"eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde\": container with ID starting with eeb0723db10943e3f98fa0fca42a811b8355d316971b4a1dfe8c1b3ee0ad5bde not found: ID does not exist" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.190512 4945 scope.go:117] "RemoveContainer" containerID="74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480" Dec 06 07:07:09 crc kubenswrapper[4945]: E1206 07:07:09.191128 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480\": container with ID starting with 74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480 not found: ID does not exist" containerID="74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.191205 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480"} err="failed to get container status \"74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480\": rpc error: code = NotFound desc = could not find container \"74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480\": container with ID starting with 74c73e816814684f079f986e9bff3a4ae8df26e518a0df2c25bbdb18e2946480 not found: ID does not exist" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.191251 4945 scope.go:117] "RemoveContainer" containerID="845576fa1982d7e18bba49498a66f9811779a237678acfa18ccbe96601cf2b7b" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.213002 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dbee646-d4c4-4ae1-baf1-36736ff75c9e" (UID: "2dbee646-d4c4-4ae1-baf1-36736ff75c9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.261955 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.262001 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2nhf\" (UniqueName: \"kubernetes.io/projected/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-kube-api-access-t2nhf\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.262018 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbee646-d4c4-4ae1-baf1-36736ff75c9e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.460533 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zqfs9"] Dec 06 07:07:09 crc kubenswrapper[4945]: I1206 07:07:09.464719 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zqfs9"] Dec 06 07:07:10 crc kubenswrapper[4945]: I1206 07:07:10.961558 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" path="/var/lib/kubelet/pods/2dbee646-d4c4-4ae1-baf1-36736ff75c9e/volumes" Dec 06 07:07:45 crc kubenswrapper[4945]: I1206 07:07:45.941199 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mlqp5"] Dec 06 07:07:45 crc kubenswrapper[4945]: I1206 07:07:45.942690 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovn-controller" containerID="cri-o://4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259" gracePeriod=30 Dec 06 07:07:45 crc kubenswrapper[4945]: I1206 07:07:45.942752 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="nbdb" containerID="cri-o://583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059" gracePeriod=30 Dec 06 07:07:45 crc kubenswrapper[4945]: I1206 07:07:45.942758 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="kube-rbac-proxy-node" containerID="cri-o://2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c" gracePeriod=30 Dec 06 07:07:45 crc kubenswrapper[4945]: I1206 07:07:45.942837 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="northd" containerID="cri-o://bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af" gracePeriod=30 Dec 06 07:07:45 crc kubenswrapper[4945]: I1206 07:07:45.942767 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee" gracePeriod=30 Dec 06 07:07:45 crc kubenswrapper[4945]: I1206 07:07:45.942871 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovn-acl-logging" containerID="cri-o://96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6" gracePeriod=30 Dec 06 07:07:45 crc kubenswrapper[4945]: I1206 07:07:45.942883 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="sbdb" containerID="cri-o://e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc" gracePeriod=30 Dec 06 07:07:45 crc kubenswrapper[4945]: I1206 07:07:45.982816 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" containerID="cri-o://d7d68c21448dc3cdea139233e91a91ca289f3040b698a3d09456305979ab0585" gracePeriod=30 Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.367149 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6v6gj_c9fc852c-2d5d-444c-b470-2728bb3dcdfd/kube-multus/2.log" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.367945 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6v6gj_c9fc852c-2d5d-444c-b470-2728bb3dcdfd/kube-multus/1.log" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.368022 4945 generic.go:334] "Generic (PLEG): container finished" podID="c9fc852c-2d5d-444c-b470-2728bb3dcdfd" containerID="f45f16e4b54f593b1b291ff367b569587c385448071b36e1380fe53e9c3f6411" exitCode=2 Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.368155 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6v6gj" event={"ID":"c9fc852c-2d5d-444c-b470-2728bb3dcdfd","Type":"ContainerDied","Data":"f45f16e4b54f593b1b291ff367b569587c385448071b36e1380fe53e9c3f6411"} Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.368229 4945 scope.go:117] "RemoveContainer" containerID="ca391eea471883373e325a7691efd4f359972d2b296084dc82973f6aa31ce0f4" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.369006 4945 scope.go:117] "RemoveContainer" containerID="f45f16e4b54f593b1b291ff367b569587c385448071b36e1380fe53e9c3f6411" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.371736 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovnkube-controller/3.log" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.375149 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovn-acl-logging/0.log" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376298 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovn-controller/0.log" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376772 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="d7d68c21448dc3cdea139233e91a91ca289f3040b698a3d09456305979ab0585" exitCode=0 Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376815 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc" exitCode=0 Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376826 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059" exitCode=0 Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376836 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af" exitCode=0 Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376846 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee" exitCode=0 Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376854 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c" exitCode=0 Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376863 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6" exitCode=143 Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376855 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"d7d68c21448dc3cdea139233e91a91ca289f3040b698a3d09456305979ab0585"} Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376916 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc"} Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376934 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059"} Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376949 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af"} Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376874 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerID="4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259" exitCode=143 Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.376966 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee"} Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.377025 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c"} Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.377049 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6"} Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.377067 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259"} Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.444118 4945 scope.go:117] "RemoveContainer" containerID="31aab9eee18160fed59305453d70b7ac3d88c67f698ade0e8b4ba21a69cb3a47" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.700471 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovn-acl-logging/0.log" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.701035 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovn-controller/0.log" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.704461 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.765947 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zghtn"] Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766197 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovn-acl-logging" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766212 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovn-acl-logging" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766223 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerName="registry-server" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766230 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerName="registry-server" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766238 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="sbdb" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766246 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="sbdb" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766260 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="kube-rbac-proxy-node" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766267 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="kube-rbac-proxy-node" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766295 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766305 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766312 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="nbdb" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766318 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="nbdb" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766327 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766333 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766342 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="northd" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766349 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="northd" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766356 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="kubecfg-setup" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766362 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="kubecfg-setup" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766369 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerName="extract-content" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766374 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerName="extract-content" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766384 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766390 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766401 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766407 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766416 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerName="extract-utilities" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766421 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerName="extract-utilities" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766432 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766437 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766445 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovn-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766451 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovn-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766567 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766582 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766593 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766608 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dbee646-d4c4-4ae1-baf1-36736ff75c9e" containerName="registry-server" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766620 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="northd" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766632 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="sbdb" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766641 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovn-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766650 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovn-acl-logging" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766659 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="kube-rbac-proxy-node" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766666 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766675 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="nbdb" Dec 06 07:07:46 crc kubenswrapper[4945]: E1206 07:07:46.766802 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766817 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766947 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.766959 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" containerName="ovnkube-controller" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.768627 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847455 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovn-node-metrics-cert\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847506 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-etc-openvswitch\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847534 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847554 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-ovn\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847580 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-log-socket\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847602 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-script-lib\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847623 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-node-log\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847649 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-bin\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847666 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-netns\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847685 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-slash\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847704 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-systemd-units\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847736 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-config\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847756 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-ovn-kubernetes\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847775 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-var-lib-openvswitch\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847803 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qh8t\" (UniqueName: \"kubernetes.io/projected/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-kube-api-access-5qh8t\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847819 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-kubelet\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847835 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-netd\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847854 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-openvswitch\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847873 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-env-overrides\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.847891 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-systemd\") pod \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\" (UID: \"4d69e33c-ba3b-4fca-a77a-a098db9d24c7\") " Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848010 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-systemd-units\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848028 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-cni-netd\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848050 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/91c03570-79d4-4f67-8366-e1e583751235-ovn-node-metrics-cert\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848076 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-run-openvswitch\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848095 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-kubelet\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848111 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-run-systemd\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848133 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-cni-bin\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848148 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/91c03570-79d4-4f67-8366-e1e583751235-ovnkube-config\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848169 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-log-socket\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848193 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-etc-openvswitch\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848214 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-node-log\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848235 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78rm7\" (UniqueName: \"kubernetes.io/projected/91c03570-79d4-4f67-8366-e1e583751235-kube-api-access-78rm7\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848261 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-slash\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848298 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-var-lib-openvswitch\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848315 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/91c03570-79d4-4f67-8366-e1e583751235-env-overrides\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848345 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-run-ovn\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848370 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-run-netns\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848389 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/91c03570-79d4-4f67-8366-e1e583751235-ovnkube-script-lib\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848405 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848431 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-run-ovn-kubernetes\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848424 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848510 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848937 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.848977 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849067 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-log-socket" (OuterVolumeSpecName: "log-socket") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849160 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849195 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849225 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849232 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849271 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-node-log" (OuterVolumeSpecName: "node-log") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849380 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849395 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849410 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849435 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-slash" (OuterVolumeSpecName: "host-slash") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849775 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849860 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.849900 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.855716 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.856102 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-kube-api-access-5qh8t" (OuterVolumeSpecName: "kube-api-access-5qh8t") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "kube-api-access-5qh8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.864686 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "4d69e33c-ba3b-4fca-a77a-a098db9d24c7" (UID: "4d69e33c-ba3b-4fca-a77a-a098db9d24c7"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.949685 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-log-socket\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.949769 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-etc-openvswitch\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.949791 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-node-log\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.949811 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78rm7\" (UniqueName: \"kubernetes.io/projected/91c03570-79d4-4f67-8366-e1e583751235-kube-api-access-78rm7\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.949838 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-slash\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.949950 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-node-log\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.949988 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/91c03570-79d4-4f67-8366-e1e583751235-env-overrides\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950010 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-var-lib-openvswitch\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.949914 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-etc-openvswitch\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.949904 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-log-socket\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950083 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-run-ovn\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950126 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-var-lib-openvswitch\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950032 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-run-ovn\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950232 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-run-netns\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950312 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/91c03570-79d4-4f67-8366-e1e583751235-ovnkube-script-lib\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950354 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950431 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-run-ovn-kubernetes\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950496 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-cni-netd\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950523 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-systemd-units\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950551 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/91c03570-79d4-4f67-8366-e1e583751235-ovn-node-metrics-cert\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950579 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-run-openvswitch\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950585 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-slash\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950615 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-kubelet\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950640 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/91c03570-79d4-4f67-8366-e1e583751235-env-overrides\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950669 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-kubelet\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950685 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-cni-netd\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950705 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950723 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-run-systemd\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950739 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-systemd-units\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950740 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-run-ovn-kubernetes\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950779 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-run-systemd\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950785 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-run-netns\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950847 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-cni-bin\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950864 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-run-openvswitch\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.950888 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/91c03570-79d4-4f67-8366-e1e583751235-host-cni-bin\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951047 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/91c03570-79d4-4f67-8366-e1e583751235-ovnkube-config\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951160 4945 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951175 4945 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951177 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/91c03570-79d4-4f67-8366-e1e583751235-ovnkube-script-lib\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951187 4945 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-slash\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951199 4945 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951211 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951223 4945 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951238 4945 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951249 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qh8t\" (UniqueName: \"kubernetes.io/projected/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-kube-api-access-5qh8t\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951260 4945 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951271 4945 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951311 4945 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951322 4945 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951332 4945 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951344 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951357 4945 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951372 4945 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951385 4945 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951399 4945 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-log-socket\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951411 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951423 4945 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d69e33c-ba3b-4fca-a77a-a098db9d24c7-node-log\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.951705 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/91c03570-79d4-4f67-8366-e1e583751235-ovnkube-config\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.955234 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/91c03570-79d4-4f67-8366-e1e583751235-ovn-node-metrics-cert\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:46 crc kubenswrapper[4945]: I1206 07:07:46.966996 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78rm7\" (UniqueName: \"kubernetes.io/projected/91c03570-79d4-4f67-8366-e1e583751235-kube-api-access-78rm7\") pod \"ovnkube-node-zghtn\" (UID: \"91c03570-79d4-4f67-8366-e1e583751235\") " pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.084942 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:47 crc kubenswrapper[4945]: W1206 07:07:47.106582 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91c03570_79d4_4f67_8366_e1e583751235.slice/crio-a9456038906c05ddb6c575f1987dd763fb2563f107704db150272adb551982bf WatchSource:0}: Error finding container a9456038906c05ddb6c575f1987dd763fb2563f107704db150272adb551982bf: Status 404 returned error can't find the container with id a9456038906c05ddb6c575f1987dd763fb2563f107704db150272adb551982bf Dec 06 07:07:47 crc kubenswrapper[4945]: E1206 07:07:47.375690 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91c03570_79d4_4f67_8366_e1e583751235.slice/crio-conmon-515a014435485f8083dfd821e1fc82e57969b9d9df71bbe09b54d63b11dcf19d.scope\": RecentStats: unable to find data in memory cache]" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.389013 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovn-acl-logging/0.log" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.389508 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mlqp5_4d69e33c-ba3b-4fca-a77a-a098db9d24c7/ovn-controller/0.log" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.390021 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" event={"ID":"4d69e33c-ba3b-4fca-a77a-a098db9d24c7","Type":"ContainerDied","Data":"eba08d28493cc703b338f067dae3837d840455078335644aa955264435c755db"} Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.390081 4945 scope.go:117] "RemoveContainer" containerID="d7d68c21448dc3cdea139233e91a91ca289f3040b698a3d09456305979ab0585" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.390119 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mlqp5" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.394304 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6v6gj_c9fc852c-2d5d-444c-b470-2728bb3dcdfd/kube-multus/2.log" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.394427 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6v6gj" event={"ID":"c9fc852c-2d5d-444c-b470-2728bb3dcdfd","Type":"ContainerStarted","Data":"a20dfc0fa18491421fd41039d2b6eaddfc08a5d815e6e7acc8b0181224e5b708"} Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.396890 4945 generic.go:334] "Generic (PLEG): container finished" podID="91c03570-79d4-4f67-8366-e1e583751235" containerID="515a014435485f8083dfd821e1fc82e57969b9d9df71bbe09b54d63b11dcf19d" exitCode=0 Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.396927 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" event={"ID":"91c03570-79d4-4f67-8366-e1e583751235","Type":"ContainerDied","Data":"515a014435485f8083dfd821e1fc82e57969b9d9df71bbe09b54d63b11dcf19d"} Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.396952 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" event={"ID":"91c03570-79d4-4f67-8366-e1e583751235","Type":"ContainerStarted","Data":"a9456038906c05ddb6c575f1987dd763fb2563f107704db150272adb551982bf"} Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.420271 4945 scope.go:117] "RemoveContainer" containerID="e8ea68286a9a4c747fb5d384632cc0eb65684ed3940ae6618c8f380fa2fde1fc" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.421706 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mlqp5"] Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.426973 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mlqp5"] Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.443587 4945 scope.go:117] "RemoveContainer" containerID="583729ca788bc002c16fe1c90f6db403bfba2f28b5378e7060882371d0e9b059" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.462462 4945 scope.go:117] "RemoveContainer" containerID="bcc8f4ff119dbd660be438577e3b95cbc802061eee9535376fc8305a5b4328af" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.494790 4945 scope.go:117] "RemoveContainer" containerID="3e46a09a0276f82a1676d6a3eeee4b66a6b93fb6d803f4a61e1c5e144e91efee" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.519105 4945 scope.go:117] "RemoveContainer" containerID="2811d9e49fbc34fd19f0a4a3a89846adc9a9516348cde371b61bde22bff6086c" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.547625 4945 scope.go:117] "RemoveContainer" containerID="96a7277cf6ea2b6fae2dbacbcb3cb350c2ed3f386ca0296080439a3db91b9da6" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.567475 4945 scope.go:117] "RemoveContainer" containerID="4afc4bfb3c54daf7e8be5bf2aa1f739d852d8efd49dffaa2c2d53eaefbd59259" Dec 06 07:07:47 crc kubenswrapper[4945]: I1206 07:07:47.585364 4945 scope.go:117] "RemoveContainer" containerID="a5b3b7bce30457469227f381c2ea91e5f355d757f2e1d5113c2528788ac6091d" Dec 06 07:07:48 crc kubenswrapper[4945]: I1206 07:07:48.406049 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" event={"ID":"91c03570-79d4-4f67-8366-e1e583751235","Type":"ContainerStarted","Data":"d50ad77cdccd394ab152068e1bda72c60ff06ab8ef4d3dcdcd0c49acb71dfd74"} Dec 06 07:07:48 crc kubenswrapper[4945]: I1206 07:07:48.406698 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" event={"ID":"91c03570-79d4-4f67-8366-e1e583751235","Type":"ContainerStarted","Data":"e3d69619943367fe34cda756abae84d24fe26cc900db1a1edd9f2d788e743d84"} Dec 06 07:07:48 crc kubenswrapper[4945]: I1206 07:07:48.406718 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" event={"ID":"91c03570-79d4-4f67-8366-e1e583751235","Type":"ContainerStarted","Data":"77a7e9e1c7e18823357a944097699eb4bc86178505e03557c17226be230d5c47"} Dec 06 07:07:48 crc kubenswrapper[4945]: I1206 07:07:48.406733 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" event={"ID":"91c03570-79d4-4f67-8366-e1e583751235","Type":"ContainerStarted","Data":"9f2143604a6abf0e1bb2b8a0b3768385f333f35af4845203deb7b59b4869c4de"} Dec 06 07:07:48 crc kubenswrapper[4945]: I1206 07:07:48.406744 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" event={"ID":"91c03570-79d4-4f67-8366-e1e583751235","Type":"ContainerStarted","Data":"6b5178b9d73d90bcf8d1c08af604b3f94f457eb2b0f651ca8f874bc00baec929"} Dec 06 07:07:48 crc kubenswrapper[4945]: I1206 07:07:48.406760 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" event={"ID":"91c03570-79d4-4f67-8366-e1e583751235","Type":"ContainerStarted","Data":"3a1db2847e24b78c50f787e3196d454fd869a1eda214f2088e9519a3dcab7cc1"} Dec 06 07:07:48 crc kubenswrapper[4945]: I1206 07:07:48.974854 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d69e33c-ba3b-4fca-a77a-a098db9d24c7" path="/var/lib/kubelet/pods/4d69e33c-ba3b-4fca-a77a-a098db9d24c7/volumes" Dec 06 07:07:51 crc kubenswrapper[4945]: I1206 07:07:51.430399 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" event={"ID":"91c03570-79d4-4f67-8366-e1e583751235","Type":"ContainerStarted","Data":"b8ab662fc1a153b9c65462b8f5a3c11928d976c100f5616942c94d7f28286dd7"} Dec 06 07:07:52 crc kubenswrapper[4945]: I1206 07:07:52.861742 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-85qdr"] Dec 06 07:07:52 crc kubenswrapper[4945]: I1206 07:07:52.863382 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:52 crc kubenswrapper[4945]: I1206 07:07:52.865236 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 06 07:07:52 crc kubenswrapper[4945]: I1206 07:07:52.865356 4945 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-mqlkm" Dec 06 07:07:52 crc kubenswrapper[4945]: I1206 07:07:52.866250 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 06 07:07:52 crc kubenswrapper[4945]: I1206 07:07:52.867020 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 06 07:07:53 crc kubenswrapper[4945]: I1206 07:07:53.031185 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/42d31853-9fa8-4d5d-85fc-db2179129bfe-crc-storage\") pod \"crc-storage-crc-85qdr\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: I1206 07:07:53.031289 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/42d31853-9fa8-4d5d-85fc-db2179129bfe-node-mnt\") pod \"crc-storage-crc-85qdr\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: I1206 07:07:53.031336 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jx6z\" (UniqueName: \"kubernetes.io/projected/42d31853-9fa8-4d5d-85fc-db2179129bfe-kube-api-access-8jx6z\") pod \"crc-storage-crc-85qdr\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: I1206 07:07:53.132805 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/42d31853-9fa8-4d5d-85fc-db2179129bfe-crc-storage\") pod \"crc-storage-crc-85qdr\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: I1206 07:07:53.133565 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/42d31853-9fa8-4d5d-85fc-db2179129bfe-node-mnt\") pod \"crc-storage-crc-85qdr\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: I1206 07:07:53.133709 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jx6z\" (UniqueName: \"kubernetes.io/projected/42d31853-9fa8-4d5d-85fc-db2179129bfe-kube-api-access-8jx6z\") pod \"crc-storage-crc-85qdr\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: I1206 07:07:53.133701 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/42d31853-9fa8-4d5d-85fc-db2179129bfe-crc-storage\") pod \"crc-storage-crc-85qdr\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: I1206 07:07:53.133898 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/42d31853-9fa8-4d5d-85fc-db2179129bfe-node-mnt\") pod \"crc-storage-crc-85qdr\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: I1206 07:07:53.157161 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jx6z\" (UniqueName: \"kubernetes.io/projected/42d31853-9fa8-4d5d-85fc-db2179129bfe-kube-api-access-8jx6z\") pod \"crc-storage-crc-85qdr\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: I1206 07:07:53.179524 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: E1206 07:07:53.206919 4945 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-85qdr_crc-storage_42d31853-9fa8-4d5d-85fc-db2179129bfe_0(a5f4ecd86af7cd743234bd7335d9f8d16246cf215e8332bba015d161c7edf91c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 07:07:53 crc kubenswrapper[4945]: E1206 07:07:53.207029 4945 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-85qdr_crc-storage_42d31853-9fa8-4d5d-85fc-db2179129bfe_0(a5f4ecd86af7cd743234bd7335d9f8d16246cf215e8332bba015d161c7edf91c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: E1206 07:07:53.207059 4945 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-85qdr_crc-storage_42d31853-9fa8-4d5d-85fc-db2179129bfe_0(a5f4ecd86af7cd743234bd7335d9f8d16246cf215e8332bba015d161c7edf91c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:53 crc kubenswrapper[4945]: E1206 07:07:53.207134 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-85qdr_crc-storage(42d31853-9fa8-4d5d-85fc-db2179129bfe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-85qdr_crc-storage(42d31853-9fa8-4d5d-85fc-db2179129bfe)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-85qdr_crc-storage_42d31853-9fa8-4d5d-85fc-db2179129bfe_0(a5f4ecd86af7cd743234bd7335d9f8d16246cf215e8332bba015d161c7edf91c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-85qdr" podUID="42d31853-9fa8-4d5d-85fc-db2179129bfe" Dec 06 07:07:55 crc kubenswrapper[4945]: I1206 07:07:55.456530 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" event={"ID":"91c03570-79d4-4f67-8366-e1e583751235","Type":"ContainerStarted","Data":"60f03832d943478f1384cf9eb7c9af43817e6989d328ae642537f42f7a2506e4"} Dec 06 07:07:55 crc kubenswrapper[4945]: I1206 07:07:55.456893 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:55 crc kubenswrapper[4945]: I1206 07:07:55.456933 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:55 crc kubenswrapper[4945]: I1206 07:07:55.485715 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:55 crc kubenswrapper[4945]: I1206 07:07:55.488093 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" podStartSLOduration=9.488075937 podStartE2EDuration="9.488075937s" podCreationTimestamp="2025-12-06 07:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:07:55.484580333 +0000 UTC m=+888.939441397" watchObservedRunningTime="2025-12-06 07:07:55.488075937 +0000 UTC m=+888.942936981" Dec 06 07:07:55 crc kubenswrapper[4945]: I1206 07:07:55.801531 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-85qdr"] Dec 06 07:07:55 crc kubenswrapper[4945]: I1206 07:07:55.801651 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:55 crc kubenswrapper[4945]: I1206 07:07:55.802028 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:55 crc kubenswrapper[4945]: E1206 07:07:55.838879 4945 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-85qdr_crc-storage_42d31853-9fa8-4d5d-85fc-db2179129bfe_0(3eacf6e84484fcb0fd391673f8f6b47a29effab3b6251db1401f1ba5563e4440): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 07:07:55 crc kubenswrapper[4945]: E1206 07:07:55.838964 4945 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-85qdr_crc-storage_42d31853-9fa8-4d5d-85fc-db2179129bfe_0(3eacf6e84484fcb0fd391673f8f6b47a29effab3b6251db1401f1ba5563e4440): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:55 crc kubenswrapper[4945]: E1206 07:07:55.838986 4945 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-85qdr_crc-storage_42d31853-9fa8-4d5d-85fc-db2179129bfe_0(3eacf6e84484fcb0fd391673f8f6b47a29effab3b6251db1401f1ba5563e4440): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:07:55 crc kubenswrapper[4945]: E1206 07:07:55.839036 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-85qdr_crc-storage(42d31853-9fa8-4d5d-85fc-db2179129bfe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-85qdr_crc-storage(42d31853-9fa8-4d5d-85fc-db2179129bfe)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-85qdr_crc-storage_42d31853-9fa8-4d5d-85fc-db2179129bfe_0(3eacf6e84484fcb0fd391673f8f6b47a29effab3b6251db1401f1ba5563e4440): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-85qdr" podUID="42d31853-9fa8-4d5d-85fc-db2179129bfe" Dec 06 07:07:56 crc kubenswrapper[4945]: I1206 07:07:56.462218 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:07:56 crc kubenswrapper[4945]: I1206 07:07:56.490968 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.545878 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lwk9d"] Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.548067 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.559027 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lwk9d"] Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.666239 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp62m\" (UniqueName: \"kubernetes.io/projected/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-kube-api-access-zp62m\") pod \"redhat-marketplace-lwk9d\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.666326 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-utilities\") pod \"redhat-marketplace-lwk9d\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.666749 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-catalog-content\") pod \"redhat-marketplace-lwk9d\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.768507 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-catalog-content\") pod \"redhat-marketplace-lwk9d\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.768648 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp62m\" (UniqueName: \"kubernetes.io/projected/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-kube-api-access-zp62m\") pod \"redhat-marketplace-lwk9d\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.769065 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-catalog-content\") pod \"redhat-marketplace-lwk9d\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.769104 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-utilities\") pod \"redhat-marketplace-lwk9d\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.769527 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-utilities\") pod \"redhat-marketplace-lwk9d\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.791400 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp62m\" (UniqueName: \"kubernetes.io/projected/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-kube-api-access-zp62m\") pod \"redhat-marketplace-lwk9d\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:02 crc kubenswrapper[4945]: I1206 07:08:02.867322 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:03 crc kubenswrapper[4945]: I1206 07:08:03.088226 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lwk9d"] Dec 06 07:08:03 crc kubenswrapper[4945]: W1206 07:08:03.095889 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7d110e5_1682_4b6f_ab2c_2f8abb5cd136.slice/crio-6b7c7d4e339cb239ae4606281565d1d942a228dc0e5988c06e93cd45e3310358 WatchSource:0}: Error finding container 6b7c7d4e339cb239ae4606281565d1d942a228dc0e5988c06e93cd45e3310358: Status 404 returned error can't find the container with id 6b7c7d4e339cb239ae4606281565d1d942a228dc0e5988c06e93cd45e3310358 Dec 06 07:08:03 crc kubenswrapper[4945]: I1206 07:08:03.501235 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwk9d" event={"ID":"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136","Type":"ContainerStarted","Data":"6b7c7d4e339cb239ae4606281565d1d942a228dc0e5988c06e93cd45e3310358"} Dec 06 07:08:04 crc kubenswrapper[4945]: I1206 07:08:04.507829 4945 generic.go:334] "Generic (PLEG): container finished" podID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerID="2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492" exitCode=0 Dec 06 07:08:04 crc kubenswrapper[4945]: I1206 07:08:04.507883 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwk9d" event={"ID":"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136","Type":"ContainerDied","Data":"2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492"} Dec 06 07:08:05 crc kubenswrapper[4945]: I1206 07:08:05.517201 4945 generic.go:334] "Generic (PLEG): container finished" podID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerID="0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee" exitCode=0 Dec 06 07:08:05 crc kubenswrapper[4945]: I1206 07:08:05.517321 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwk9d" event={"ID":"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136","Type":"ContainerDied","Data":"0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee"} Dec 06 07:08:06 crc kubenswrapper[4945]: I1206 07:08:06.527649 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwk9d" event={"ID":"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136","Type":"ContainerStarted","Data":"fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc"} Dec 06 07:08:06 crc kubenswrapper[4945]: I1206 07:08:06.550538 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lwk9d" podStartSLOduration=3.128507806 podStartE2EDuration="4.550516787s" podCreationTimestamp="2025-12-06 07:08:02 +0000 UTC" firstStartedPulling="2025-12-06 07:08:04.509938996 +0000 UTC m=+897.964800060" lastFinishedPulling="2025-12-06 07:08:05.931948007 +0000 UTC m=+899.386809041" observedRunningTime="2025-12-06 07:08:06.547692561 +0000 UTC m=+900.002553635" watchObservedRunningTime="2025-12-06 07:08:06.550516787 +0000 UTC m=+900.005377831" Dec 06 07:08:09 crc kubenswrapper[4945]: I1206 07:08:09.952271 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:08:09 crc kubenswrapper[4945]: I1206 07:08:09.953087 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:08:10 crc kubenswrapper[4945]: I1206 07:08:10.125184 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-85qdr"] Dec 06 07:08:10 crc kubenswrapper[4945]: W1206 07:08:10.132118 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42d31853_9fa8_4d5d_85fc_db2179129bfe.slice/crio-d186fac5cfb9b52d2cd378b2cbd1efb85e03eb906860101dc67486e01ff66b35 WatchSource:0}: Error finding container d186fac5cfb9b52d2cd378b2cbd1efb85e03eb906860101dc67486e01ff66b35: Status 404 returned error can't find the container with id d186fac5cfb9b52d2cd378b2cbd1efb85e03eb906860101dc67486e01ff66b35 Dec 06 07:08:10 crc kubenswrapper[4945]: I1206 07:08:10.548914 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-85qdr" event={"ID":"42d31853-9fa8-4d5d-85fc-db2179129bfe","Type":"ContainerStarted","Data":"d186fac5cfb9b52d2cd378b2cbd1efb85e03eb906860101dc67486e01ff66b35"} Dec 06 07:08:12 crc kubenswrapper[4945]: I1206 07:08:12.563671 4945 generic.go:334] "Generic (PLEG): container finished" podID="42d31853-9fa8-4d5d-85fc-db2179129bfe" containerID="d1a103b3b7982b4586c200a020c46be4223b800dabf858bae432a976d96a0efd" exitCode=0 Dec 06 07:08:12 crc kubenswrapper[4945]: I1206 07:08:12.563799 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-85qdr" event={"ID":"42d31853-9fa8-4d5d-85fc-db2179129bfe","Type":"ContainerDied","Data":"d1a103b3b7982b4586c200a020c46be4223b800dabf858bae432a976d96a0efd"} Dec 06 07:08:12 crc kubenswrapper[4945]: I1206 07:08:12.867826 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:12 crc kubenswrapper[4945]: I1206 07:08:12.867908 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:12 crc kubenswrapper[4945]: I1206 07:08:12.909038 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:13 crc kubenswrapper[4945]: I1206 07:08:13.612474 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:13 crc kubenswrapper[4945]: I1206 07:08:13.660253 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lwk9d"] Dec 06 07:08:13 crc kubenswrapper[4945]: I1206 07:08:13.776452 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:08:13 crc kubenswrapper[4945]: I1206 07:08:13.918658 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jx6z\" (UniqueName: \"kubernetes.io/projected/42d31853-9fa8-4d5d-85fc-db2179129bfe-kube-api-access-8jx6z\") pod \"42d31853-9fa8-4d5d-85fc-db2179129bfe\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " Dec 06 07:08:13 crc kubenswrapper[4945]: I1206 07:08:13.918727 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/42d31853-9fa8-4d5d-85fc-db2179129bfe-crc-storage\") pod \"42d31853-9fa8-4d5d-85fc-db2179129bfe\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " Dec 06 07:08:13 crc kubenswrapper[4945]: I1206 07:08:13.918833 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/42d31853-9fa8-4d5d-85fc-db2179129bfe-node-mnt\") pod \"42d31853-9fa8-4d5d-85fc-db2179129bfe\" (UID: \"42d31853-9fa8-4d5d-85fc-db2179129bfe\") " Dec 06 07:08:13 crc kubenswrapper[4945]: I1206 07:08:13.919436 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/42d31853-9fa8-4d5d-85fc-db2179129bfe-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "42d31853-9fa8-4d5d-85fc-db2179129bfe" (UID: "42d31853-9fa8-4d5d-85fc-db2179129bfe"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:08:13 crc kubenswrapper[4945]: I1206 07:08:13.925164 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42d31853-9fa8-4d5d-85fc-db2179129bfe-kube-api-access-8jx6z" (OuterVolumeSpecName: "kube-api-access-8jx6z") pod "42d31853-9fa8-4d5d-85fc-db2179129bfe" (UID: "42d31853-9fa8-4d5d-85fc-db2179129bfe"). InnerVolumeSpecName "kube-api-access-8jx6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:08:13 crc kubenswrapper[4945]: I1206 07:08:13.934265 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42d31853-9fa8-4d5d-85fc-db2179129bfe-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "42d31853-9fa8-4d5d-85fc-db2179129bfe" (UID: "42d31853-9fa8-4d5d-85fc-db2179129bfe"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:08:14 crc kubenswrapper[4945]: I1206 07:08:14.021201 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jx6z\" (UniqueName: \"kubernetes.io/projected/42d31853-9fa8-4d5d-85fc-db2179129bfe-kube-api-access-8jx6z\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:14 crc kubenswrapper[4945]: I1206 07:08:14.021270 4945 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/42d31853-9fa8-4d5d-85fc-db2179129bfe-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:14 crc kubenswrapper[4945]: I1206 07:08:14.021303 4945 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/42d31853-9fa8-4d5d-85fc-db2179129bfe-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:14 crc kubenswrapper[4945]: I1206 07:08:14.576434 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-85qdr" Dec 06 07:08:14 crc kubenswrapper[4945]: I1206 07:08:14.576464 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-85qdr" event={"ID":"42d31853-9fa8-4d5d-85fc-db2179129bfe","Type":"ContainerDied","Data":"d186fac5cfb9b52d2cd378b2cbd1efb85e03eb906860101dc67486e01ff66b35"} Dec 06 07:08:14 crc kubenswrapper[4945]: I1206 07:08:14.576556 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d186fac5cfb9b52d2cd378b2cbd1efb85e03eb906860101dc67486e01ff66b35" Dec 06 07:08:15 crc kubenswrapper[4945]: I1206 07:08:15.581856 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lwk9d" podUID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerName="registry-server" containerID="cri-o://fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc" gracePeriod=2 Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.410220 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.556692 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-catalog-content\") pod \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.556806 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-utilities\") pod \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.556920 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp62m\" (UniqueName: \"kubernetes.io/projected/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-kube-api-access-zp62m\") pod \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\" (UID: \"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136\") " Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.557708 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-utilities" (OuterVolumeSpecName: "utilities") pod "a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" (UID: "a7d110e5-1682-4b6f-ab2c-2f8abb5cd136"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.561830 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-kube-api-access-zp62m" (OuterVolumeSpecName: "kube-api-access-zp62m") pod "a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" (UID: "a7d110e5-1682-4b6f-ab2c-2f8abb5cd136"). InnerVolumeSpecName "kube-api-access-zp62m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.578590 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" (UID: "a7d110e5-1682-4b6f-ab2c-2f8abb5cd136"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.590371 4945 generic.go:334] "Generic (PLEG): container finished" podID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerID="fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc" exitCode=0 Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.590417 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwk9d" event={"ID":"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136","Type":"ContainerDied","Data":"fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc"} Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.590428 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lwk9d" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.590447 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwk9d" event={"ID":"a7d110e5-1682-4b6f-ab2c-2f8abb5cd136","Type":"ContainerDied","Data":"6b7c7d4e339cb239ae4606281565d1d942a228dc0e5988c06e93cd45e3310358"} Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.590470 4945 scope.go:117] "RemoveContainer" containerID="fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.610072 4945 scope.go:117] "RemoveContainer" containerID="0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.626416 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lwk9d"] Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.629787 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lwk9d"] Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.646750 4945 scope.go:117] "RemoveContainer" containerID="2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.657970 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.658015 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp62m\" (UniqueName: \"kubernetes.io/projected/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-kube-api-access-zp62m\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.658026 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.660368 4945 scope.go:117] "RemoveContainer" containerID="fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc" Dec 06 07:08:16 crc kubenswrapper[4945]: E1206 07:08:16.660868 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc\": container with ID starting with fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc not found: ID does not exist" containerID="fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.660899 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc"} err="failed to get container status \"fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc\": rpc error: code = NotFound desc = could not find container \"fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc\": container with ID starting with fa2225fe29e8136bd7f240722a3d94d72ea7ad2b8725f2906e11566c6db90adc not found: ID does not exist" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.660921 4945 scope.go:117] "RemoveContainer" containerID="0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee" Dec 06 07:08:16 crc kubenswrapper[4945]: E1206 07:08:16.661154 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee\": container with ID starting with 0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee not found: ID does not exist" containerID="0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.661182 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee"} err="failed to get container status \"0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee\": rpc error: code = NotFound desc = could not find container \"0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee\": container with ID starting with 0ff09720f178d00da24e0c2e34b132da6394c1444bf5e6f7c5692820d79d86ee not found: ID does not exist" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.661200 4945 scope.go:117] "RemoveContainer" containerID="2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492" Dec 06 07:08:16 crc kubenswrapper[4945]: E1206 07:08:16.661420 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492\": container with ID starting with 2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492 not found: ID does not exist" containerID="2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.661447 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492"} err="failed to get container status \"2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492\": rpc error: code = NotFound desc = could not find container \"2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492\": container with ID starting with 2a5589684d9ccaa7f89b736c17f406e998ed501751d12472d5401c20b01da492 not found: ID does not exist" Dec 06 07:08:16 crc kubenswrapper[4945]: I1206 07:08:16.961214 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" path="/var/lib/kubelet/pods/a7d110e5-1682-4b6f-ab2c-2f8abb5cd136/volumes" Dec 06 07:08:17 crc kubenswrapper[4945]: I1206 07:08:17.111649 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zghtn" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.079430 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd"] Dec 06 07:08:21 crc kubenswrapper[4945]: E1206 07:08:21.080315 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerName="registry-server" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.080335 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerName="registry-server" Dec 06 07:08:21 crc kubenswrapper[4945]: E1206 07:08:21.080356 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d31853-9fa8-4d5d-85fc-db2179129bfe" containerName="storage" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.080365 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d31853-9fa8-4d5d-85fc-db2179129bfe" containerName="storage" Dec 06 07:08:21 crc kubenswrapper[4945]: E1206 07:08:21.080386 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerName="extract-content" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.080395 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerName="extract-content" Dec 06 07:08:21 crc kubenswrapper[4945]: E1206 07:08:21.080410 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerName="extract-utilities" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.080418 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerName="extract-utilities" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.080551 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d110e5-1682-4b6f-ab2c-2f8abb5cd136" containerName="registry-server" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.080568 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d31853-9fa8-4d5d-85fc-db2179129bfe" containerName="storage" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.081501 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.084251 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.089676 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd"] Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.212108 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.212188 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25j48\" (UniqueName: \"kubernetes.io/projected/466623ca-b9e9-4bf2-82e9-9049583349b7-kube-api-access-25j48\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.212537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.314331 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.314414 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.314444 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25j48\" (UniqueName: \"kubernetes.io/projected/466623ca-b9e9-4bf2-82e9-9049583349b7-kube-api-access-25j48\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.315175 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.315365 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.332657 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25j48\" (UniqueName: \"kubernetes.io/projected/466623ca-b9e9-4bf2-82e9-9049583349b7-kube-api-access-25j48\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.398307 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.587750 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd"] Dec 06 07:08:21 crc kubenswrapper[4945]: I1206 07:08:21.617813 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" event={"ID":"466623ca-b9e9-4bf2-82e9-9049583349b7","Type":"ContainerStarted","Data":"3f7238033ecff7c63b6570cb6fff29c3fd4642834111380e401125e4bfbbf91e"} Dec 06 07:08:22 crc kubenswrapper[4945]: I1206 07:08:22.625595 4945 generic.go:334] "Generic (PLEG): container finished" podID="466623ca-b9e9-4bf2-82e9-9049583349b7" containerID="df8393dc32847c689a47bfde86df76a21ac7c260107f1a243ce1d4aec24ffa3a" exitCode=0 Dec 06 07:08:22 crc kubenswrapper[4945]: I1206 07:08:22.625654 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" event={"ID":"466623ca-b9e9-4bf2-82e9-9049583349b7","Type":"ContainerDied","Data":"df8393dc32847c689a47bfde86df76a21ac7c260107f1a243ce1d4aec24ffa3a"} Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.445021 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vhc8m"] Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.446609 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.458964 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vhc8m"] Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.545008 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-utilities\") pod \"redhat-operators-vhc8m\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.545057 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-catalog-content\") pod \"redhat-operators-vhc8m\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.545094 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw4s6\" (UniqueName: \"kubernetes.io/projected/3c745ada-5516-4eb4-b288-8b26e21d5f17-kube-api-access-gw4s6\") pod \"redhat-operators-vhc8m\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.646165 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw4s6\" (UniqueName: \"kubernetes.io/projected/3c745ada-5516-4eb4-b288-8b26e21d5f17-kube-api-access-gw4s6\") pod \"redhat-operators-vhc8m\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.646265 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-utilities\") pod \"redhat-operators-vhc8m\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.646321 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-catalog-content\") pod \"redhat-operators-vhc8m\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.646840 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-catalog-content\") pod \"redhat-operators-vhc8m\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.646868 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-utilities\") pod \"redhat-operators-vhc8m\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.676300 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw4s6\" (UniqueName: \"kubernetes.io/projected/3c745ada-5516-4eb4-b288-8b26e21d5f17-kube-api-access-gw4s6\") pod \"redhat-operators-vhc8m\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:23 crc kubenswrapper[4945]: I1206 07:08:23.770990 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:24 crc kubenswrapper[4945]: I1206 07:08:23.999731 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vhc8m"] Dec 06 07:08:24 crc kubenswrapper[4945]: I1206 07:08:24.641576 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerID="0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef" exitCode=0 Dec 06 07:08:24 crc kubenswrapper[4945]: I1206 07:08:24.641639 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhc8m" event={"ID":"3c745ada-5516-4eb4-b288-8b26e21d5f17","Type":"ContainerDied","Data":"0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef"} Dec 06 07:08:24 crc kubenswrapper[4945]: I1206 07:08:24.641952 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhc8m" event={"ID":"3c745ada-5516-4eb4-b288-8b26e21d5f17","Type":"ContainerStarted","Data":"9a66b0f46e070f1a9652d0d9e4a2498e1b6c55567efbc274d3f58912792dfef9"} Dec 06 07:08:24 crc kubenswrapper[4945]: I1206 07:08:24.644226 4945 generic.go:334] "Generic (PLEG): container finished" podID="466623ca-b9e9-4bf2-82e9-9049583349b7" containerID="7c84440e6a20bf53628ac90cf0ab068ebd2fcbeb0c34821ecd17753ad8d4d349" exitCode=0 Dec 06 07:08:24 crc kubenswrapper[4945]: I1206 07:08:24.644306 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" event={"ID":"466623ca-b9e9-4bf2-82e9-9049583349b7","Type":"ContainerDied","Data":"7c84440e6a20bf53628ac90cf0ab068ebd2fcbeb0c34821ecd17753ad8d4d349"} Dec 06 07:08:25 crc kubenswrapper[4945]: I1206 07:08:25.654623 4945 generic.go:334] "Generic (PLEG): container finished" podID="466623ca-b9e9-4bf2-82e9-9049583349b7" containerID="4be1a6402cad9ad3d341249b5066bdab5d5afa0ebf820c833685008a71857dd3" exitCode=0 Dec 06 07:08:25 crc kubenswrapper[4945]: I1206 07:08:25.654670 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" event={"ID":"466623ca-b9e9-4bf2-82e9-9049583349b7","Type":"ContainerDied","Data":"4be1a6402cad9ad3d341249b5066bdab5d5afa0ebf820c833685008a71857dd3"} Dec 06 07:08:26 crc kubenswrapper[4945]: I1206 07:08:26.850538 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:26 crc kubenswrapper[4945]: I1206 07:08:26.989185 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25j48\" (UniqueName: \"kubernetes.io/projected/466623ca-b9e9-4bf2-82e9-9049583349b7-kube-api-access-25j48\") pod \"466623ca-b9e9-4bf2-82e9-9049583349b7\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " Dec 06 07:08:26 crc kubenswrapper[4945]: I1206 07:08:26.989241 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-util\") pod \"466623ca-b9e9-4bf2-82e9-9049583349b7\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " Dec 06 07:08:26 crc kubenswrapper[4945]: I1206 07:08:26.989389 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-bundle\") pod \"466623ca-b9e9-4bf2-82e9-9049583349b7\" (UID: \"466623ca-b9e9-4bf2-82e9-9049583349b7\") " Dec 06 07:08:26 crc kubenswrapper[4945]: I1206 07:08:26.990130 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-bundle" (OuterVolumeSpecName: "bundle") pod "466623ca-b9e9-4bf2-82e9-9049583349b7" (UID: "466623ca-b9e9-4bf2-82e9-9049583349b7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:08:26 crc kubenswrapper[4945]: I1206 07:08:26.994234 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/466623ca-b9e9-4bf2-82e9-9049583349b7-kube-api-access-25j48" (OuterVolumeSpecName: "kube-api-access-25j48") pod "466623ca-b9e9-4bf2-82e9-9049583349b7" (UID: "466623ca-b9e9-4bf2-82e9-9049583349b7"). InnerVolumeSpecName "kube-api-access-25j48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:08:27 crc kubenswrapper[4945]: I1206 07:08:27.090597 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:27 crc kubenswrapper[4945]: I1206 07:08:27.090636 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25j48\" (UniqueName: \"kubernetes.io/projected/466623ca-b9e9-4bf2-82e9-9049583349b7-kube-api-access-25j48\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:27 crc kubenswrapper[4945]: I1206 07:08:27.669527 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" event={"ID":"466623ca-b9e9-4bf2-82e9-9049583349b7","Type":"ContainerDied","Data":"3f7238033ecff7c63b6570cb6fff29c3fd4642834111380e401125e4bfbbf91e"} Dec 06 07:08:27 crc kubenswrapper[4945]: I1206 07:08:27.669578 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f7238033ecff7c63b6570cb6fff29c3fd4642834111380e401125e4bfbbf91e" Dec 06 07:08:27 crc kubenswrapper[4945]: I1206 07:08:27.669604 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.637710 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2"] Dec 06 07:08:29 crc kubenswrapper[4945]: E1206 07:08:29.638232 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="466623ca-b9e9-4bf2-82e9-9049583349b7" containerName="pull" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.638250 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="466623ca-b9e9-4bf2-82e9-9049583349b7" containerName="pull" Dec 06 07:08:29 crc kubenswrapper[4945]: E1206 07:08:29.638272 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="466623ca-b9e9-4bf2-82e9-9049583349b7" containerName="util" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.638300 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="466623ca-b9e9-4bf2-82e9-9049583349b7" containerName="util" Dec 06 07:08:29 crc kubenswrapper[4945]: E1206 07:08:29.638309 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="466623ca-b9e9-4bf2-82e9-9049583349b7" containerName="extract" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.638315 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="466623ca-b9e9-4bf2-82e9-9049583349b7" containerName="extract" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.638398 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="466623ca-b9e9-4bf2-82e9-9049583349b7" containerName="extract" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.638781 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.641032 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.641312 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.641424 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bzj7l" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.655500 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2"] Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.823086 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqnbb\" (UniqueName: \"kubernetes.io/projected/bfdca761-893a-4fb8-b500-af0d0bdf44e8-kube-api-access-lqnbb\") pod \"nmstate-operator-5b5b58f5c8-s72f2\" (UID: \"bfdca761-893a-4fb8-b500-af0d0bdf44e8\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.924644 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqnbb\" (UniqueName: \"kubernetes.io/projected/bfdca761-893a-4fb8-b500-af0d0bdf44e8-kube-api-access-lqnbb\") pod \"nmstate-operator-5b5b58f5c8-s72f2\" (UID: \"bfdca761-893a-4fb8-b500-af0d0bdf44e8\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.942920 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqnbb\" (UniqueName: \"kubernetes.io/projected/bfdca761-893a-4fb8-b500-af0d0bdf44e8-kube-api-access-lqnbb\") pod \"nmstate-operator-5b5b58f5c8-s72f2\" (UID: \"bfdca761-893a-4fb8-b500-af0d0bdf44e8\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2" Dec 06 07:08:29 crc kubenswrapper[4945]: I1206 07:08:29.952026 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2" Dec 06 07:08:30 crc kubenswrapper[4945]: I1206 07:08:30.127190 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2"] Dec 06 07:08:30 crc kubenswrapper[4945]: I1206 07:08:30.370479 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-util" (OuterVolumeSpecName: "util") pod "466623ca-b9e9-4bf2-82e9-9049583349b7" (UID: "466623ca-b9e9-4bf2-82e9-9049583349b7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:08:30 crc kubenswrapper[4945]: I1206 07:08:30.430463 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/466623ca-b9e9-4bf2-82e9-9049583349b7-util\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:30 crc kubenswrapper[4945]: I1206 07:08:30.687365 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2" event={"ID":"bfdca761-893a-4fb8-b500-af0d0bdf44e8","Type":"ContainerStarted","Data":"767695df3287e39e3c35159352fc5e66f84293abb3b4f88339779814b2b65546"} Dec 06 07:08:31 crc kubenswrapper[4945]: I1206 07:08:31.696618 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhc8m" event={"ID":"3c745ada-5516-4eb4-b288-8b26e21d5f17","Type":"ContainerStarted","Data":"a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e"} Dec 06 07:08:33 crc kubenswrapper[4945]: I1206 07:08:33.858740 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerID="a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e" exitCode=0 Dec 06 07:08:33 crc kubenswrapper[4945]: I1206 07:08:33.858876 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhc8m" event={"ID":"3c745ada-5516-4eb4-b288-8b26e21d5f17","Type":"ContainerDied","Data":"a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e"} Dec 06 07:08:37 crc kubenswrapper[4945]: I1206 07:08:37.160625 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2" event={"ID":"bfdca761-893a-4fb8-b500-af0d0bdf44e8","Type":"ContainerStarted","Data":"bbe9bf6df6c30d8ade5986e1496428bbfcc3446e3f959e4af61fc03af8b54207"} Dec 06 07:08:37 crc kubenswrapper[4945]: I1206 07:08:37.164488 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhc8m" event={"ID":"3c745ada-5516-4eb4-b288-8b26e21d5f17","Type":"ContainerStarted","Data":"b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99"} Dec 06 07:08:37 crc kubenswrapper[4945]: I1206 07:08:37.181084 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s72f2" podStartSLOduration=1.804373958 podStartE2EDuration="8.18106128s" podCreationTimestamp="2025-12-06 07:08:29 +0000 UTC" firstStartedPulling="2025-12-06 07:08:30.134824267 +0000 UTC m=+923.589685311" lastFinishedPulling="2025-12-06 07:08:36.511511589 +0000 UTC m=+929.966372633" observedRunningTime="2025-12-06 07:08:37.175152801 +0000 UTC m=+930.630013845" watchObservedRunningTime="2025-12-06 07:08:37.18106128 +0000 UTC m=+930.635922324" Dec 06 07:08:37 crc kubenswrapper[4945]: I1206 07:08:37.199140 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vhc8m" podStartSLOduration=2.003313815 podStartE2EDuration="14.199116615s" podCreationTimestamp="2025-12-06 07:08:23 +0000 UTC" firstStartedPulling="2025-12-06 07:08:24.643951504 +0000 UTC m=+918.098812548" lastFinishedPulling="2025-12-06 07:08:36.839754304 +0000 UTC m=+930.294615348" observedRunningTime="2025-12-06 07:08:37.1963161 +0000 UTC m=+930.651177154" watchObservedRunningTime="2025-12-06 07:08:37.199116615 +0000 UTC m=+930.653977669" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.145055 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-4999s"] Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.145926 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4999s" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.148979 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-97n7h" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.154593 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq"] Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.156071 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" Dec 06 07:08:38 crc kubenswrapper[4945]: W1206 07:08:38.157804 4945 reflector.go:561] object-"openshift-nmstate"/"openshift-nmstate-webhook": failed to list *v1.Secret: secrets "openshift-nmstate-webhook" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Dec 06 07:08:38 crc kubenswrapper[4945]: E1206 07:08:38.157857 4945 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"openshift-nmstate-webhook\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-nmstate-webhook\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.158369 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-4999s"] Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.171319 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-lb98q"] Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.172159 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.194872 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq"] Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.237852 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqv8z\" (UniqueName: \"kubernetes.io/projected/9fc802d9-93bc-46af-acba-93258bfb4890-kube-api-access-xqv8z\") pod \"nmstate-webhook-5f6d4c5ccb-4kdsq\" (UID: \"9fc802d9-93bc-46af-acba-93258bfb4890\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.237919 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzqgp\" (UniqueName: \"kubernetes.io/projected/fcf83bdc-5289-4109-af7f-10c3198eac36-kube-api-access-gzqgp\") pod \"nmstate-metrics-7f946cbc9-4999s\" (UID: \"fcf83bdc-5289-4109-af7f-10c3198eac36\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4999s" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.237948 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9fc802d9-93bc-46af-acba-93258bfb4890-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4kdsq\" (UID: \"9fc802d9-93bc-46af-acba-93258bfb4890\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.339380 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ae31a742-6313-4d6c-bfe2-989f1e46d401-dbus-socket\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.339435 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9fc802d9-93bc-46af-acba-93258bfb4890-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4kdsq\" (UID: \"9fc802d9-93bc-46af-acba-93258bfb4890\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.339464 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ae31a742-6313-4d6c-bfe2-989f1e46d401-ovs-socket\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.339509 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ae31a742-6313-4d6c-bfe2-989f1e46d401-nmstate-lock\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.339531 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqv8z\" (UniqueName: \"kubernetes.io/projected/9fc802d9-93bc-46af-acba-93258bfb4890-kube-api-access-xqv8z\") pod \"nmstate-webhook-5f6d4c5ccb-4kdsq\" (UID: \"9fc802d9-93bc-46af-acba-93258bfb4890\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.339558 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pf8q\" (UniqueName: \"kubernetes.io/projected/ae31a742-6313-4d6c-bfe2-989f1e46d401-kube-api-access-9pf8q\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.339610 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzqgp\" (UniqueName: \"kubernetes.io/projected/fcf83bdc-5289-4109-af7f-10c3198eac36-kube-api-access-gzqgp\") pod \"nmstate-metrics-7f946cbc9-4999s\" (UID: \"fcf83bdc-5289-4109-af7f-10c3198eac36\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4999s" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.365440 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq"] Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.366327 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.381939 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.382247 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-nj2nv" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.386685 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.393211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqv8z\" (UniqueName: \"kubernetes.io/projected/9fc802d9-93bc-46af-acba-93258bfb4890-kube-api-access-xqv8z\") pod \"nmstate-webhook-5f6d4c5ccb-4kdsq\" (UID: \"9fc802d9-93bc-46af-acba-93258bfb4890\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.395119 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzqgp\" (UniqueName: \"kubernetes.io/projected/fcf83bdc-5289-4109-af7f-10c3198eac36-kube-api-access-gzqgp\") pod \"nmstate-metrics-7f946cbc9-4999s\" (UID: \"fcf83bdc-5289-4109-af7f-10c3198eac36\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4999s" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.406726 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq"] Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.440883 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ae31a742-6313-4d6c-bfe2-989f1e46d401-dbus-socket\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.440965 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ae31a742-6313-4d6c-bfe2-989f1e46d401-ovs-socket\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.441002 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ae31a742-6313-4d6c-bfe2-989f1e46d401-nmstate-lock\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.441032 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pf8q\" (UniqueName: \"kubernetes.io/projected/ae31a742-6313-4d6c-bfe2-989f1e46d401-kube-api-access-9pf8q\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.441361 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ae31a742-6313-4d6c-bfe2-989f1e46d401-nmstate-lock\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.441372 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ae31a742-6313-4d6c-bfe2-989f1e46d401-ovs-socket\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.441724 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ae31a742-6313-4d6c-bfe2-989f1e46d401-dbus-socket\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.464808 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pf8q\" (UniqueName: \"kubernetes.io/projected/ae31a742-6313-4d6c-bfe2-989f1e46d401-kube-api-access-9pf8q\") pod \"nmstate-handler-lb98q\" (UID: \"ae31a742-6313-4d6c-bfe2-989f1e46d401\") " pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.469031 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4999s" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.508663 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.542522 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/49085c7f-2b2f-4692-a76a-b1a159f8e121-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-lrsrq\" (UID: \"49085c7f-2b2f-4692-a76a-b1a159f8e121\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.542870 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dnpk\" (UniqueName: \"kubernetes.io/projected/49085c7f-2b2f-4692-a76a-b1a159f8e121-kube-api-access-5dnpk\") pod \"nmstate-console-plugin-7fbb5f6569-lrsrq\" (UID: \"49085c7f-2b2f-4692-a76a-b1a159f8e121\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.542953 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/49085c7f-2b2f-4692-a76a-b1a159f8e121-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-lrsrq\" (UID: \"49085c7f-2b2f-4692-a76a-b1a159f8e121\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.585666 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7cb7bdc4d7-j4kp8"] Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.586351 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.602548 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cb7bdc4d7-j4kp8"] Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.643956 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/49085c7f-2b2f-4692-a76a-b1a159f8e121-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-lrsrq\" (UID: \"49085c7f-2b2f-4692-a76a-b1a159f8e121\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.644041 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dnpk\" (UniqueName: \"kubernetes.io/projected/49085c7f-2b2f-4692-a76a-b1a159f8e121-kube-api-access-5dnpk\") pod \"nmstate-console-plugin-7fbb5f6569-lrsrq\" (UID: \"49085c7f-2b2f-4692-a76a-b1a159f8e121\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.644145 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/49085c7f-2b2f-4692-a76a-b1a159f8e121-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-lrsrq\" (UID: \"49085c7f-2b2f-4692-a76a-b1a159f8e121\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:38 crc kubenswrapper[4945]: E1206 07:08:38.644412 4945 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 06 07:08:38 crc kubenswrapper[4945]: E1206 07:08:38.644481 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49085c7f-2b2f-4692-a76a-b1a159f8e121-plugin-serving-cert podName:49085c7f-2b2f-4692-a76a-b1a159f8e121 nodeName:}" failed. No retries permitted until 2025-12-06 07:08:39.14445542 +0000 UTC m=+932.599316464 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/49085c7f-2b2f-4692-a76a-b1a159f8e121-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-lrsrq" (UID: "49085c7f-2b2f-4692-a76a-b1a159f8e121") : secret "plugin-serving-cert" not found Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.645548 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/49085c7f-2b2f-4692-a76a-b1a159f8e121-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-lrsrq\" (UID: \"49085c7f-2b2f-4692-a76a-b1a159f8e121\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.665481 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dnpk\" (UniqueName: \"kubernetes.io/projected/49085c7f-2b2f-4692-a76a-b1a159f8e121-kube-api-access-5dnpk\") pod \"nmstate-console-plugin-7fbb5f6569-lrsrq\" (UID: \"49085c7f-2b2f-4692-a76a-b1a159f8e121\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.745657 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-console-serving-cert\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.745719 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-oauth-serving-cert\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.745746 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-trusted-ca-bundle\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.745790 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lvcv\" (UniqueName: \"kubernetes.io/projected/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-kube-api-access-8lvcv\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.745806 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-console-oauth-config\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.745848 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-console-config\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.745864 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-service-ca\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.846892 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-console-serving-cert\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.846992 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-oauth-serving-cert\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.847039 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-trusted-ca-bundle\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.847076 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lvcv\" (UniqueName: \"kubernetes.io/projected/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-kube-api-access-8lvcv\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.847092 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-console-oauth-config\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.847128 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-console-config\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.847143 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-service-ca\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.848422 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-service-ca\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.848624 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-oauth-serving-cert\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.849169 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-console-config\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:38 crc kubenswrapper[4945]: I1206 07:08:38.849586 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-trusted-ca-bundle\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:38.852397 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-console-serving-cert\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:38.874751 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-console-oauth-config\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:38.876355 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lvcv\" (UniqueName: \"kubernetes.io/projected/6775e1cc-01a7-4c67-aae6-528c8d23f8f6-kube-api-access-8lvcv\") pod \"console-7cb7bdc4d7-j4kp8\" (UID: \"6775e1cc-01a7-4c67-aae6-528c8d23f8f6\") " pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:38.905317 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.157535 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/49085c7f-2b2f-4692-a76a-b1a159f8e121-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-lrsrq\" (UID: \"49085c7f-2b2f-4692-a76a-b1a159f8e121\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.162165 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/49085c7f-2b2f-4692-a76a-b1a159f8e121-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-lrsrq\" (UID: \"49085c7f-2b2f-4692-a76a-b1a159f8e121\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.182087 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lb98q" event={"ID":"ae31a742-6313-4d6c-bfe2-989f1e46d401","Type":"ContainerStarted","Data":"fa5275fd51db0db61d8cf6989f1bb376adf3beec2f6c1f19abebeaf63b6107f0"} Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.182793 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-4999s"] Dec 06 07:08:39 crc kubenswrapper[4945]: W1206 07:08:39.198203 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcf83bdc_5289_4109_af7f_10c3198eac36.slice/crio-b27dc89d26bd695f6659be96cd9f1b09f3fcb7e0f2a9dcd13c8f5cd140245e94 WatchSource:0}: Error finding container b27dc89d26bd695f6659be96cd9f1b09f3fcb7e0f2a9dcd13c8f5cd140245e94: Status 404 returned error can't find the container with id b27dc89d26bd695f6659be96cd9f1b09f3fcb7e0f2a9dcd13c8f5cd140245e94 Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.287370 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.299554 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cb7bdc4d7-j4kp8"] Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.303069 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 06 07:08:39 crc kubenswrapper[4945]: W1206 07:08:39.314874 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6775e1cc_01a7_4c67_aae6_528c8d23f8f6.slice/crio-4ef012f108e35a5c85d6b98f51eea87f237e6bc38ed7289208f1c9c4f219b4c1 WatchSource:0}: Error finding container 4ef012f108e35a5c85d6b98f51eea87f237e6bc38ed7289208f1c9c4f219b4c1: Status 404 returned error can't find the container with id 4ef012f108e35a5c85d6b98f51eea87f237e6bc38ed7289208f1c9c4f219b4c1 Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.315934 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9fc802d9-93bc-46af-acba-93258bfb4890-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4kdsq\" (UID: \"9fc802d9-93bc-46af-acba-93258bfb4890\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.377266 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.515453 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq"] Dec 06 07:08:39 crc kubenswrapper[4945]: W1206 07:08:39.524214 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49085c7f_2b2f_4692_a76a_b1a159f8e121.slice/crio-310898f2d72dc0216e80e310d15d12ecc13cf98190503c023cf3e9a71972f23f WatchSource:0}: Error finding container 310898f2d72dc0216e80e310d15d12ecc13cf98190503c023cf3e9a71972f23f: Status 404 returned error can't find the container with id 310898f2d72dc0216e80e310d15d12ecc13cf98190503c023cf3e9a71972f23f Dec 06 07:08:39 crc kubenswrapper[4945]: I1206 07:08:39.613713 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq"] Dec 06 07:08:40 crc kubenswrapper[4945]: I1206 07:08:40.188615 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" event={"ID":"9fc802d9-93bc-46af-acba-93258bfb4890","Type":"ContainerStarted","Data":"e180473c9b705bda5de6dafc1ea40033f1c0e39f56b542f6a1839b03964e35fa"} Dec 06 07:08:40 crc kubenswrapper[4945]: I1206 07:08:40.189958 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" event={"ID":"49085c7f-2b2f-4692-a76a-b1a159f8e121","Type":"ContainerStarted","Data":"310898f2d72dc0216e80e310d15d12ecc13cf98190503c023cf3e9a71972f23f"} Dec 06 07:08:40 crc kubenswrapper[4945]: I1206 07:08:40.190943 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4999s" event={"ID":"fcf83bdc-5289-4109-af7f-10c3198eac36","Type":"ContainerStarted","Data":"b27dc89d26bd695f6659be96cd9f1b09f3fcb7e0f2a9dcd13c8f5cd140245e94"} Dec 06 07:08:40 crc kubenswrapper[4945]: I1206 07:08:40.191756 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cb7bdc4d7-j4kp8" event={"ID":"6775e1cc-01a7-4c67-aae6-528c8d23f8f6","Type":"ContainerStarted","Data":"4ef012f108e35a5c85d6b98f51eea87f237e6bc38ed7289208f1c9c4f219b4c1"} Dec 06 07:08:41 crc kubenswrapper[4945]: I1206 07:08:41.200131 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cb7bdc4d7-j4kp8" event={"ID":"6775e1cc-01a7-4c67-aae6-528c8d23f8f6","Type":"ContainerStarted","Data":"2afd4e02263f2db5707741155ceb00f4d4c9d9e937e37f3680738fd0fc6fa11a"} Dec 06 07:08:41 crc kubenswrapper[4945]: I1206 07:08:41.221807 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7cb7bdc4d7-j4kp8" podStartSLOduration=3.221785196 podStartE2EDuration="3.221785196s" podCreationTimestamp="2025-12-06 07:08:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:08:41.217932792 +0000 UTC m=+934.672793836" watchObservedRunningTime="2025-12-06 07:08:41.221785196 +0000 UTC m=+934.676646240" Dec 06 07:08:43 crc kubenswrapper[4945]: I1206 07:08:43.771873 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:43 crc kubenswrapper[4945]: I1206 07:08:43.771919 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:43 crc kubenswrapper[4945]: I1206 07:08:43.809879 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:44 crc kubenswrapper[4945]: I1206 07:08:44.266884 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:44 crc kubenswrapper[4945]: I1206 07:08:44.316164 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vhc8m"] Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.244663 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" event={"ID":"49085c7f-2b2f-4692-a76a-b1a159f8e121","Type":"ContainerStarted","Data":"f809ea672f3675c82bf89f6d3014695eeea1f662c2013bdb32800842e2c3a42f"} Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.246750 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lb98q" event={"ID":"ae31a742-6313-4d6c-bfe2-989f1e46d401","Type":"ContainerStarted","Data":"f153878de0bfeb11983e5ac4844932198ad8b176ceffd218b3b5c74a3015282e"} Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.246930 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.248816 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4999s" event={"ID":"fcf83bdc-5289-4109-af7f-10c3198eac36","Type":"ContainerStarted","Data":"0e09efcffcc94c95e2b8fb801293985050ba5c2de7b045fdfebae0d28603acbe"} Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.252317 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" event={"ID":"9fc802d9-93bc-46af-acba-93258bfb4890","Type":"ContainerStarted","Data":"4f48494116e3228faeaae64164465b403c23d88a976aabe6cc91ff64f79a83e1"} Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.252468 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vhc8m" podUID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerName="registry-server" containerID="cri-o://b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99" gracePeriod=2 Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.268070 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lrsrq" podStartSLOduration=1.924855767 podStartE2EDuration="8.268045638s" podCreationTimestamp="2025-12-06 07:08:38 +0000 UTC" firstStartedPulling="2025-12-06 07:08:39.52692472 +0000 UTC m=+932.981785774" lastFinishedPulling="2025-12-06 07:08:45.870114601 +0000 UTC m=+939.324975645" observedRunningTime="2025-12-06 07:08:46.265189941 +0000 UTC m=+939.720050985" watchObservedRunningTime="2025-12-06 07:08:46.268045638 +0000 UTC m=+939.722906682" Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.291796 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" podStartSLOduration=2.037956994 podStartE2EDuration="8.291758585s" podCreationTimestamp="2025-12-06 07:08:38 +0000 UTC" firstStartedPulling="2025-12-06 07:08:39.624828659 +0000 UTC m=+933.079689713" lastFinishedPulling="2025-12-06 07:08:45.87863026 +0000 UTC m=+939.333491304" observedRunningTime="2025-12-06 07:08:46.282151606 +0000 UTC m=+939.737012650" watchObservedRunningTime="2025-12-06 07:08:46.291758585 +0000 UTC m=+939.746619629" Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.304016 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-lb98q" podStartSLOduration=0.974947046 podStartE2EDuration="8.303991413s" podCreationTimestamp="2025-12-06 07:08:38 +0000 UTC" firstStartedPulling="2025-12-06 07:08:38.539661036 +0000 UTC m=+931.994522080" lastFinishedPulling="2025-12-06 07:08:45.868705403 +0000 UTC m=+939.323566447" observedRunningTime="2025-12-06 07:08:46.300751386 +0000 UTC m=+939.755612440" watchObservedRunningTime="2025-12-06 07:08:46.303991413 +0000 UTC m=+939.758852457" Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.574645 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.664107 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-utilities\") pod \"3c745ada-5516-4eb4-b288-8b26e21d5f17\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.664363 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-catalog-content\") pod \"3c745ada-5516-4eb4-b288-8b26e21d5f17\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.664404 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw4s6\" (UniqueName: \"kubernetes.io/projected/3c745ada-5516-4eb4-b288-8b26e21d5f17-kube-api-access-gw4s6\") pod \"3c745ada-5516-4eb4-b288-8b26e21d5f17\" (UID: \"3c745ada-5516-4eb4-b288-8b26e21d5f17\") " Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.665081 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-utilities" (OuterVolumeSpecName: "utilities") pod "3c745ada-5516-4eb4-b288-8b26e21d5f17" (UID: "3c745ada-5516-4eb4-b288-8b26e21d5f17"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.670040 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c745ada-5516-4eb4-b288-8b26e21d5f17-kube-api-access-gw4s6" (OuterVolumeSpecName: "kube-api-access-gw4s6") pod "3c745ada-5516-4eb4-b288-8b26e21d5f17" (UID: "3c745ada-5516-4eb4-b288-8b26e21d5f17"). InnerVolumeSpecName "kube-api-access-gw4s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.766749 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.766809 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw4s6\" (UniqueName: \"kubernetes.io/projected/3c745ada-5516-4eb4-b288-8b26e21d5f17-kube-api-access-gw4s6\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.787206 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c745ada-5516-4eb4-b288-8b26e21d5f17" (UID: "3c745ada-5516-4eb4-b288-8b26e21d5f17"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:08:46 crc kubenswrapper[4945]: I1206 07:08:46.869189 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c745ada-5516-4eb4-b288-8b26e21d5f17-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.262905 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerID="b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99" exitCode=0 Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.262953 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhc8m" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.262970 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhc8m" event={"ID":"3c745ada-5516-4eb4-b288-8b26e21d5f17","Type":"ContainerDied","Data":"b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99"} Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.264073 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhc8m" event={"ID":"3c745ada-5516-4eb4-b288-8b26e21d5f17","Type":"ContainerDied","Data":"9a66b0f46e070f1a9652d0d9e4a2498e1b6c55567efbc274d3f58912792dfef9"} Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.264137 4945 scope.go:117] "RemoveContainer" containerID="b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.264595 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.289320 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vhc8m"] Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.291703 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vhc8m"] Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.308486 4945 scope.go:117] "RemoveContainer" containerID="a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.330697 4945 scope.go:117] "RemoveContainer" containerID="0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.349659 4945 scope.go:117] "RemoveContainer" containerID="b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99" Dec 06 07:08:47 crc kubenswrapper[4945]: E1206 07:08:47.350480 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99\": container with ID starting with b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99 not found: ID does not exist" containerID="b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.350526 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99"} err="failed to get container status \"b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99\": rpc error: code = NotFound desc = could not find container \"b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99\": container with ID starting with b9b360920f83783861fe6cac0b6ada74b9cbda86b2a476cd48936610cc254f99 not found: ID does not exist" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.350562 4945 scope.go:117] "RemoveContainer" containerID="a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e" Dec 06 07:08:47 crc kubenswrapper[4945]: E1206 07:08:47.351043 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e\": container with ID starting with a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e not found: ID does not exist" containerID="a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.351085 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e"} err="failed to get container status \"a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e\": rpc error: code = NotFound desc = could not find container \"a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e\": container with ID starting with a7ea8914229a29f68c633ad3a1b6fd5223810ef40c1e4725314d811f6c396a4e not found: ID does not exist" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.351107 4945 scope.go:117] "RemoveContainer" containerID="0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef" Dec 06 07:08:47 crc kubenswrapper[4945]: E1206 07:08:47.351694 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef\": container with ID starting with 0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef not found: ID does not exist" containerID="0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef" Dec 06 07:08:47 crc kubenswrapper[4945]: I1206 07:08:47.351728 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef"} err="failed to get container status \"0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef\": rpc error: code = NotFound desc = could not find container \"0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef\": container with ID starting with 0595a5976d2dc687bb5039c88bd4ed670e595beca5b919998b74e4ff6292d7ef not found: ID does not exist" Dec 06 07:08:48 crc kubenswrapper[4945]: I1206 07:08:48.273002 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4999s" event={"ID":"fcf83bdc-5289-4109-af7f-10c3198eac36","Type":"ContainerStarted","Data":"8e3b92d329e6e45e4a93572d78476928bcc48cb75a1ea95a2cbff4ae09a02dd4"} Dec 06 07:08:48 crc kubenswrapper[4945]: I1206 07:08:48.299752 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4999s" podStartSLOduration=1.423237814 podStartE2EDuration="10.2997115s" podCreationTimestamp="2025-12-06 07:08:38 +0000 UTC" firstStartedPulling="2025-12-06 07:08:39.199837565 +0000 UTC m=+932.654698609" lastFinishedPulling="2025-12-06 07:08:48.076311251 +0000 UTC m=+941.531172295" observedRunningTime="2025-12-06 07:08:48.293183235 +0000 UTC m=+941.748044299" watchObservedRunningTime="2025-12-06 07:08:48.2997115 +0000 UTC m=+941.754572564" Dec 06 07:08:48 crc kubenswrapper[4945]: I1206 07:08:48.906540 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:48 crc kubenswrapper[4945]: I1206 07:08:48.907012 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:48 crc kubenswrapper[4945]: I1206 07:08:48.913400 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:48 crc kubenswrapper[4945]: I1206 07:08:48.961246 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c745ada-5516-4eb4-b288-8b26e21d5f17" path="/var/lib/kubelet/pods/3c745ada-5516-4eb4-b288-8b26e21d5f17/volumes" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.283203 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7cb7bdc4d7-j4kp8" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.337861 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-j2tjw"] Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.451814 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xn8db"] Dec 06 07:08:49 crc kubenswrapper[4945]: E1206 07:08:49.452550 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerName="extract-content" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.452568 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerName="extract-content" Dec 06 07:08:49 crc kubenswrapper[4945]: E1206 07:08:49.452579 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerName="registry-server" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.452587 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerName="registry-server" Dec 06 07:08:49 crc kubenswrapper[4945]: E1206 07:08:49.452603 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerName="extract-utilities" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.452611 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerName="extract-utilities" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.452759 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c745ada-5516-4eb4-b288-8b26e21d5f17" containerName="registry-server" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.453707 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.467440 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xn8db"] Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.612792 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-utilities\") pod \"community-operators-xn8db\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.613254 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5dr6\" (UniqueName: \"kubernetes.io/projected/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-kube-api-access-z5dr6\") pod \"community-operators-xn8db\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.613444 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-catalog-content\") pod \"community-operators-xn8db\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.715127 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-catalog-content\") pod \"community-operators-xn8db\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.715601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-utilities\") pod \"community-operators-xn8db\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.715711 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5dr6\" (UniqueName: \"kubernetes.io/projected/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-kube-api-access-z5dr6\") pod \"community-operators-xn8db\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.715827 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-catalog-content\") pod \"community-operators-xn8db\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.716258 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-utilities\") pod \"community-operators-xn8db\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.744026 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5dr6\" (UniqueName: \"kubernetes.io/projected/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-kube-api-access-z5dr6\") pod \"community-operators-xn8db\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:49 crc kubenswrapper[4945]: I1206 07:08:49.779459 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:50 crc kubenswrapper[4945]: I1206 07:08:50.080163 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xn8db"] Dec 06 07:08:50 crc kubenswrapper[4945]: I1206 07:08:50.286936 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xn8db" event={"ID":"45bb0a8d-2a50-4a2c-a26a-8935b13b884a","Type":"ContainerStarted","Data":"a92050215d803d7868171a6cee94487aa2421e0110fd12200e341710762ae1c1"} Dec 06 07:08:51 crc kubenswrapper[4945]: I1206 07:08:51.297548 4945 generic.go:334] "Generic (PLEG): container finished" podID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerID="d0f1d413142bafab1d5da5f1314d100841f290b73cf8f392fd6c13fdb87f0aac" exitCode=0 Dec 06 07:08:51 crc kubenswrapper[4945]: I1206 07:08:51.297628 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xn8db" event={"ID":"45bb0a8d-2a50-4a2c-a26a-8935b13b884a","Type":"ContainerDied","Data":"d0f1d413142bafab1d5da5f1314d100841f290b73cf8f392fd6c13fdb87f0aac"} Dec 06 07:08:52 crc kubenswrapper[4945]: I1206 07:08:52.305603 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xn8db" event={"ID":"45bb0a8d-2a50-4a2c-a26a-8935b13b884a","Type":"ContainerStarted","Data":"39da0e83ad901b870de4513f15468a10a7558d9646670dcae8401c5aadc98974"} Dec 06 07:08:53 crc kubenswrapper[4945]: I1206 07:08:53.313017 4945 generic.go:334] "Generic (PLEG): container finished" podID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerID="39da0e83ad901b870de4513f15468a10a7558d9646670dcae8401c5aadc98974" exitCode=0 Dec 06 07:08:53 crc kubenswrapper[4945]: I1206 07:08:53.313074 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xn8db" event={"ID":"45bb0a8d-2a50-4a2c-a26a-8935b13b884a","Type":"ContainerDied","Data":"39da0e83ad901b870de4513f15468a10a7558d9646670dcae8401c5aadc98974"} Dec 06 07:08:53 crc kubenswrapper[4945]: I1206 07:08:53.531428 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-lb98q" Dec 06 07:08:54 crc kubenswrapper[4945]: I1206 07:08:54.320935 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xn8db" event={"ID":"45bb0a8d-2a50-4a2c-a26a-8935b13b884a","Type":"ContainerStarted","Data":"e4526becd148452b74621317ba8cc6c100130ee0c1e093e7e5f2372f63beaa21"} Dec 06 07:08:54 crc kubenswrapper[4945]: I1206 07:08:54.340970 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xn8db" podStartSLOduration=2.866326564 podStartE2EDuration="5.340948883s" podCreationTimestamp="2025-12-06 07:08:49 +0000 UTC" firstStartedPulling="2025-12-06 07:08:51.300472957 +0000 UTC m=+944.755334011" lastFinishedPulling="2025-12-06 07:08:53.775095296 +0000 UTC m=+947.229956330" observedRunningTime="2025-12-06 07:08:54.337590273 +0000 UTC m=+947.792451317" watchObservedRunningTime="2025-12-06 07:08:54.340948883 +0000 UTC m=+947.795809927" Dec 06 07:08:59 crc kubenswrapper[4945]: I1206 07:08:59.384651 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kdsq" Dec 06 07:08:59 crc kubenswrapper[4945]: I1206 07:08:59.780764 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:59 crc kubenswrapper[4945]: I1206 07:08:59.780813 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:08:59 crc kubenswrapper[4945]: I1206 07:08:59.820788 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:09:00 crc kubenswrapper[4945]: I1206 07:09:00.408025 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:09:00 crc kubenswrapper[4945]: I1206 07:09:00.449838 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xn8db"] Dec 06 07:09:02 crc kubenswrapper[4945]: I1206 07:09:02.362383 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xn8db" podUID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerName="registry-server" containerID="cri-o://e4526becd148452b74621317ba8cc6c100130ee0c1e093e7e5f2372f63beaa21" gracePeriod=2 Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.381548 4945 generic.go:334] "Generic (PLEG): container finished" podID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerID="e4526becd148452b74621317ba8cc6c100130ee0c1e093e7e5f2372f63beaa21" exitCode=0 Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.381619 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xn8db" event={"ID":"45bb0a8d-2a50-4a2c-a26a-8935b13b884a","Type":"ContainerDied","Data":"e4526becd148452b74621317ba8cc6c100130ee0c1e093e7e5f2372f63beaa21"} Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.547670 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.714008 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5dr6\" (UniqueName: \"kubernetes.io/projected/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-kube-api-access-z5dr6\") pod \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.714079 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-utilities\") pod \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.714145 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-catalog-content\") pod \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\" (UID: \"45bb0a8d-2a50-4a2c-a26a-8935b13b884a\") " Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.715331 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-utilities" (OuterVolumeSpecName: "utilities") pod "45bb0a8d-2a50-4a2c-a26a-8935b13b884a" (UID: "45bb0a8d-2a50-4a2c-a26a-8935b13b884a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.719809 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-kube-api-access-z5dr6" (OuterVolumeSpecName: "kube-api-access-z5dr6") pod "45bb0a8d-2a50-4a2c-a26a-8935b13b884a" (UID: "45bb0a8d-2a50-4a2c-a26a-8935b13b884a"). InnerVolumeSpecName "kube-api-access-z5dr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.767920 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45bb0a8d-2a50-4a2c-a26a-8935b13b884a" (UID: "45bb0a8d-2a50-4a2c-a26a-8935b13b884a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.817970 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.818030 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5dr6\" (UniqueName: \"kubernetes.io/projected/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-kube-api-access-z5dr6\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:04 crc kubenswrapper[4945]: I1206 07:09:04.818048 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45bb0a8d-2a50-4a2c-a26a-8935b13b884a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:05 crc kubenswrapper[4945]: I1206 07:09:05.391420 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xn8db" Dec 06 07:09:05 crc kubenswrapper[4945]: I1206 07:09:05.391418 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xn8db" event={"ID":"45bb0a8d-2a50-4a2c-a26a-8935b13b884a","Type":"ContainerDied","Data":"a92050215d803d7868171a6cee94487aa2421e0110fd12200e341710762ae1c1"} Dec 06 07:09:05 crc kubenswrapper[4945]: I1206 07:09:05.391576 4945 scope.go:117] "RemoveContainer" containerID="e4526becd148452b74621317ba8cc6c100130ee0c1e093e7e5f2372f63beaa21" Dec 06 07:09:05 crc kubenswrapper[4945]: I1206 07:09:05.413679 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xn8db"] Dec 06 07:09:05 crc kubenswrapper[4945]: I1206 07:09:05.418773 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xn8db"] Dec 06 07:09:05 crc kubenswrapper[4945]: I1206 07:09:05.421978 4945 scope.go:117] "RemoveContainer" containerID="39da0e83ad901b870de4513f15468a10a7558d9646670dcae8401c5aadc98974" Dec 06 07:09:05 crc kubenswrapper[4945]: I1206 07:09:05.443163 4945 scope.go:117] "RemoveContainer" containerID="d0f1d413142bafab1d5da5f1314d100841f290b73cf8f392fd6c13fdb87f0aac" Dec 06 07:09:06 crc kubenswrapper[4945]: I1206 07:09:06.960989 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" path="/var/lib/kubelet/pods/45bb0a8d-2a50-4a2c-a26a-8935b13b884a/volumes" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.042507 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx"] Dec 06 07:09:12 crc kubenswrapper[4945]: E1206 07:09:12.043323 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerName="extract-utilities" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.043341 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerName="extract-utilities" Dec 06 07:09:12 crc kubenswrapper[4945]: E1206 07:09:12.043358 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerName="registry-server" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.043366 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerName="registry-server" Dec 06 07:09:12 crc kubenswrapper[4945]: E1206 07:09:12.043378 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerName="extract-content" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.043385 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerName="extract-content" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.043505 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="45bb0a8d-2a50-4a2c-a26a-8935b13b884a" containerName="registry-server" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.044520 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.046185 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.055597 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx"] Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.121213 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.121264 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk8gh\" (UniqueName: \"kubernetes.io/projected/8168c399-bb2d-4c71-ae91-d96758be7a42-kube-api-access-bk8gh\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.121403 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.222881 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.222987 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.223035 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk8gh\" (UniqueName: \"kubernetes.io/projected/8168c399-bb2d-4c71-ae91-d96758be7a42-kube-api-access-bk8gh\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.223470 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.223745 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.249666 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk8gh\" (UniqueName: \"kubernetes.io/projected/8168c399-bb2d-4c71-ae91-d96758be7a42-kube-api-access-bk8gh\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.363660 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:12 crc kubenswrapper[4945]: I1206 07:09:12.548924 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx"] Dec 06 07:09:13 crc kubenswrapper[4945]: I1206 07:09:13.435818 4945 generic.go:334] "Generic (PLEG): container finished" podID="8168c399-bb2d-4c71-ae91-d96758be7a42" containerID="a4cf01237cd0e1404eb53375fa194b7def9314be3d498c36956284eea4fb5565" exitCode=0 Dec 06 07:09:13 crc kubenswrapper[4945]: I1206 07:09:13.435884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" event={"ID":"8168c399-bb2d-4c71-ae91-d96758be7a42","Type":"ContainerDied","Data":"a4cf01237cd0e1404eb53375fa194b7def9314be3d498c36956284eea4fb5565"} Dec 06 07:09:13 crc kubenswrapper[4945]: I1206 07:09:13.436109 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" event={"ID":"8168c399-bb2d-4c71-ae91-d96758be7a42","Type":"ContainerStarted","Data":"879802f9e90a066c83f6d80ecf6c334ad573c28d088a42e8dbb38c5c40067b89"} Dec 06 07:09:14 crc kubenswrapper[4945]: I1206 07:09:14.378891 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-j2tjw" podUID="50e16132-af33-4dbf-b4f4-a9b0c8554e24" containerName="console" containerID="cri-o://2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717" gracePeriod=15 Dec 06 07:09:14 crc kubenswrapper[4945]: I1206 07:09:14.922156 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-j2tjw_50e16132-af33-4dbf-b4f4-a9b0c8554e24/console/0.log" Dec 06 07:09:14 crc kubenswrapper[4945]: I1206 07:09:14.922495 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.063803 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-config\") pod \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.063912 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-oauth-serving-cert\") pod \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.063984 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-serving-cert\") pod \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.064022 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmfp6\" (UniqueName: \"kubernetes.io/projected/50e16132-af33-4dbf-b4f4-a9b0c8554e24-kube-api-access-rmfp6\") pod \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.064085 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-service-ca\") pod \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.064112 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-oauth-config\") pod \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.064131 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-trusted-ca-bundle\") pod \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\" (UID: \"50e16132-af33-4dbf-b4f4-a9b0c8554e24\") " Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.065158 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-service-ca" (OuterVolumeSpecName: "service-ca") pod "50e16132-af33-4dbf-b4f4-a9b0c8554e24" (UID: "50e16132-af33-4dbf-b4f4-a9b0c8554e24"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.065223 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-config" (OuterVolumeSpecName: "console-config") pod "50e16132-af33-4dbf-b4f4-a9b0c8554e24" (UID: "50e16132-af33-4dbf-b4f4-a9b0c8554e24"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.065245 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "50e16132-af33-4dbf-b4f4-a9b0c8554e24" (UID: "50e16132-af33-4dbf-b4f4-a9b0c8554e24"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.065510 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "50e16132-af33-4dbf-b4f4-a9b0c8554e24" (UID: "50e16132-af33-4dbf-b4f4-a9b0c8554e24"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.070528 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "50e16132-af33-4dbf-b4f4-a9b0c8554e24" (UID: "50e16132-af33-4dbf-b4f4-a9b0c8554e24"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.070626 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50e16132-af33-4dbf-b4f4-a9b0c8554e24-kube-api-access-rmfp6" (OuterVolumeSpecName: "kube-api-access-rmfp6") pod "50e16132-af33-4dbf-b4f4-a9b0c8554e24" (UID: "50e16132-af33-4dbf-b4f4-a9b0c8554e24"). InnerVolumeSpecName "kube-api-access-rmfp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.071208 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "50e16132-af33-4dbf-b4f4-a9b0c8554e24" (UID: "50e16132-af33-4dbf-b4f4-a9b0c8554e24"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.165163 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.165201 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.165219 4945 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.165233 4945 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.165244 4945 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/50e16132-af33-4dbf-b4f4-a9b0c8554e24-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.165257 4945 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/50e16132-af33-4dbf-b4f4-a9b0c8554e24-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.165267 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmfp6\" (UniqueName: \"kubernetes.io/projected/50e16132-af33-4dbf-b4f4-a9b0c8554e24-kube-api-access-rmfp6\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.448813 4945 generic.go:334] "Generic (PLEG): container finished" podID="8168c399-bb2d-4c71-ae91-d96758be7a42" containerID="6c6750278bb5c91a9dcc1e278ac34f1f2e7ef490235e8c93b1ff3bb82252f12c" exitCode=0 Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.448891 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" event={"ID":"8168c399-bb2d-4c71-ae91-d96758be7a42","Type":"ContainerDied","Data":"6c6750278bb5c91a9dcc1e278ac34f1f2e7ef490235e8c93b1ff3bb82252f12c"} Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.450419 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-j2tjw_50e16132-af33-4dbf-b4f4-a9b0c8554e24/console/0.log" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.450473 4945 generic.go:334] "Generic (PLEG): container finished" podID="50e16132-af33-4dbf-b4f4-a9b0c8554e24" containerID="2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717" exitCode=2 Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.450500 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-j2tjw" event={"ID":"50e16132-af33-4dbf-b4f4-a9b0c8554e24","Type":"ContainerDied","Data":"2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717"} Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.450529 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-j2tjw" event={"ID":"50e16132-af33-4dbf-b4f4-a9b0c8554e24","Type":"ContainerDied","Data":"72e210eea0aa09037ed3274ef7f7dc35f84fc3eb838760a9f0e30f787c77cb04"} Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.450548 4945 scope.go:117] "RemoveContainer" containerID="2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.450669 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-j2tjw" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.470645 4945 scope.go:117] "RemoveContainer" containerID="2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717" Dec 06 07:09:15 crc kubenswrapper[4945]: E1206 07:09:15.473469 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717\": container with ID starting with 2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717 not found: ID does not exist" containerID="2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.473516 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717"} err="failed to get container status \"2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717\": rpc error: code = NotFound desc = could not find container \"2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717\": container with ID starting with 2f3cd0dac0494e92fe24950088a13327096838f736538b84e0d2df2d61ee3717 not found: ID does not exist" Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.487972 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-j2tjw"] Dec 06 07:09:15 crc kubenswrapper[4945]: I1206 07:09:15.492659 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-j2tjw"] Dec 06 07:09:16 crc kubenswrapper[4945]: I1206 07:09:16.457086 4945 generic.go:334] "Generic (PLEG): container finished" podID="8168c399-bb2d-4c71-ae91-d96758be7a42" containerID="3201e47a74d0176e796fe9fb5293f6841a1336ca19958c301b55326ff9684a12" exitCode=0 Dec 06 07:09:16 crc kubenswrapper[4945]: I1206 07:09:16.457135 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" event={"ID":"8168c399-bb2d-4c71-ae91-d96758be7a42","Type":"ContainerDied","Data":"3201e47a74d0176e796fe9fb5293f6841a1336ca19958c301b55326ff9684a12"} Dec 06 07:09:16 crc kubenswrapper[4945]: I1206 07:09:16.961145 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50e16132-af33-4dbf-b4f4-a9b0c8554e24" path="/var/lib/kubelet/pods/50e16132-af33-4dbf-b4f4-a9b0c8554e24/volumes" Dec 06 07:09:17 crc kubenswrapper[4945]: I1206 07:09:17.700361 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:17 crc kubenswrapper[4945]: I1206 07:09:17.798373 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bk8gh\" (UniqueName: \"kubernetes.io/projected/8168c399-bb2d-4c71-ae91-d96758be7a42-kube-api-access-bk8gh\") pod \"8168c399-bb2d-4c71-ae91-d96758be7a42\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " Dec 06 07:09:17 crc kubenswrapper[4945]: I1206 07:09:17.798894 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-util\") pod \"8168c399-bb2d-4c71-ae91-d96758be7a42\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " Dec 06 07:09:17 crc kubenswrapper[4945]: I1206 07:09:17.799034 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-bundle\") pod \"8168c399-bb2d-4c71-ae91-d96758be7a42\" (UID: \"8168c399-bb2d-4c71-ae91-d96758be7a42\") " Dec 06 07:09:17 crc kubenswrapper[4945]: I1206 07:09:17.800942 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-bundle" (OuterVolumeSpecName: "bundle") pod "8168c399-bb2d-4c71-ae91-d96758be7a42" (UID: "8168c399-bb2d-4c71-ae91-d96758be7a42"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:09:17 crc kubenswrapper[4945]: I1206 07:09:17.804436 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8168c399-bb2d-4c71-ae91-d96758be7a42-kube-api-access-bk8gh" (OuterVolumeSpecName: "kube-api-access-bk8gh") pod "8168c399-bb2d-4c71-ae91-d96758be7a42" (UID: "8168c399-bb2d-4c71-ae91-d96758be7a42"). InnerVolumeSpecName "kube-api-access-bk8gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:09:17 crc kubenswrapper[4945]: I1206 07:09:17.814356 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-util" (OuterVolumeSpecName: "util") pod "8168c399-bb2d-4c71-ae91-d96758be7a42" (UID: "8168c399-bb2d-4c71-ae91-d96758be7a42"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:09:17 crc kubenswrapper[4945]: I1206 07:09:17.901139 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bk8gh\" (UniqueName: \"kubernetes.io/projected/8168c399-bb2d-4c71-ae91-d96758be7a42-kube-api-access-bk8gh\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:17 crc kubenswrapper[4945]: I1206 07:09:17.901727 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-util\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:17 crc kubenswrapper[4945]: I1206 07:09:17.901830 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8168c399-bb2d-4c71-ae91-d96758be7a42-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:09:18 crc kubenswrapper[4945]: I1206 07:09:18.474239 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" event={"ID":"8168c399-bb2d-4c71-ae91-d96758be7a42","Type":"ContainerDied","Data":"879802f9e90a066c83f6d80ecf6c334ad573c28d088a42e8dbb38c5c40067b89"} Dec 06 07:09:18 crc kubenswrapper[4945]: I1206 07:09:18.474305 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="879802f9e90a066c83f6d80ecf6c334ad573c28d088a42e8dbb38c5c40067b89" Dec 06 07:09:18 crc kubenswrapper[4945]: I1206 07:09:18.474312 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx" Dec 06 07:09:18 crc kubenswrapper[4945]: E1206 07:09:18.559148 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8168c399_bb2d_4c71_ae91_d96758be7a42.slice/crio-879802f9e90a066c83f6d80ecf6c334ad573c28d088a42e8dbb38c5c40067b89\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8168c399_bb2d_4c71_ae91_d96758be7a42.slice\": RecentStats: unable to find data in memory cache]" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.798218 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt"] Dec 06 07:09:27 crc kubenswrapper[4945]: E1206 07:09:27.798993 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8168c399-bb2d-4c71-ae91-d96758be7a42" containerName="util" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.799008 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8168c399-bb2d-4c71-ae91-d96758be7a42" containerName="util" Dec 06 07:09:27 crc kubenswrapper[4945]: E1206 07:09:27.799023 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50e16132-af33-4dbf-b4f4-a9b0c8554e24" containerName="console" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.799030 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="50e16132-af33-4dbf-b4f4-a9b0c8554e24" containerName="console" Dec 06 07:09:27 crc kubenswrapper[4945]: E1206 07:09:27.799050 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8168c399-bb2d-4c71-ae91-d96758be7a42" containerName="extract" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.799133 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8168c399-bb2d-4c71-ae91-d96758be7a42" containerName="extract" Dec 06 07:09:27 crc kubenswrapper[4945]: E1206 07:09:27.799153 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8168c399-bb2d-4c71-ae91-d96758be7a42" containerName="pull" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.799160 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8168c399-bb2d-4c71-ae91-d96758be7a42" containerName="pull" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.799301 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="50e16132-af33-4dbf-b4f4-a9b0c8554e24" containerName="console" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.799317 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8168c399-bb2d-4c71-ae91-d96758be7a42" containerName="extract" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.799797 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.802946 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.803255 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8tmsz" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.807874 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.807896 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.807931 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.812178 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt"] Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.866355 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ae4cf098-3915-418a-ae09-e1d08bfc523b-apiservice-cert\") pod \"metallb-operator-controller-manager-6dd5c5658d-plhqt\" (UID: \"ae4cf098-3915-418a-ae09-e1d08bfc523b\") " pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.866429 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7lkh\" (UniqueName: \"kubernetes.io/projected/ae4cf098-3915-418a-ae09-e1d08bfc523b-kube-api-access-k7lkh\") pod \"metallb-operator-controller-manager-6dd5c5658d-plhqt\" (UID: \"ae4cf098-3915-418a-ae09-e1d08bfc523b\") " pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.866580 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ae4cf098-3915-418a-ae09-e1d08bfc523b-webhook-cert\") pod \"metallb-operator-controller-manager-6dd5c5658d-plhqt\" (UID: \"ae4cf098-3915-418a-ae09-e1d08bfc523b\") " pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.967732 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ae4cf098-3915-418a-ae09-e1d08bfc523b-apiservice-cert\") pod \"metallb-operator-controller-manager-6dd5c5658d-plhqt\" (UID: \"ae4cf098-3915-418a-ae09-e1d08bfc523b\") " pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.967792 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7lkh\" (UniqueName: \"kubernetes.io/projected/ae4cf098-3915-418a-ae09-e1d08bfc523b-kube-api-access-k7lkh\") pod \"metallb-operator-controller-manager-6dd5c5658d-plhqt\" (UID: \"ae4cf098-3915-418a-ae09-e1d08bfc523b\") " pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.967821 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ae4cf098-3915-418a-ae09-e1d08bfc523b-webhook-cert\") pod \"metallb-operator-controller-manager-6dd5c5658d-plhqt\" (UID: \"ae4cf098-3915-418a-ae09-e1d08bfc523b\") " pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.974954 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ae4cf098-3915-418a-ae09-e1d08bfc523b-webhook-cert\") pod \"metallb-operator-controller-manager-6dd5c5658d-plhqt\" (UID: \"ae4cf098-3915-418a-ae09-e1d08bfc523b\") " pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.980577 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ae4cf098-3915-418a-ae09-e1d08bfc523b-apiservice-cert\") pod \"metallb-operator-controller-manager-6dd5c5658d-plhqt\" (UID: \"ae4cf098-3915-418a-ae09-e1d08bfc523b\") " pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:27 crc kubenswrapper[4945]: I1206 07:09:27.989631 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7lkh\" (UniqueName: \"kubernetes.io/projected/ae4cf098-3915-418a-ae09-e1d08bfc523b-kube-api-access-k7lkh\") pod \"metallb-operator-controller-manager-6dd5c5658d-plhqt\" (UID: \"ae4cf098-3915-418a-ae09-e1d08bfc523b\") " pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.118619 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.129444 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp"] Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.130267 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.134550 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.134768 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.134993 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-sqv7g" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.155056 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp"] Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.272301 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0-webhook-cert\") pod \"metallb-operator-webhook-server-685db896b9-sn7lp\" (UID: \"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0\") " pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.272715 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0-apiservice-cert\") pod \"metallb-operator-webhook-server-685db896b9-sn7lp\" (UID: \"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0\") " pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.272769 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbp7f\" (UniqueName: \"kubernetes.io/projected/30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0-kube-api-access-fbp7f\") pod \"metallb-operator-webhook-server-685db896b9-sn7lp\" (UID: \"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0\") " pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.373895 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0-apiservice-cert\") pod \"metallb-operator-webhook-server-685db896b9-sn7lp\" (UID: \"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0\") " pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.373963 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbp7f\" (UniqueName: \"kubernetes.io/projected/30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0-kube-api-access-fbp7f\") pod \"metallb-operator-webhook-server-685db896b9-sn7lp\" (UID: \"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0\") " pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.374056 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0-webhook-cert\") pod \"metallb-operator-webhook-server-685db896b9-sn7lp\" (UID: \"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0\") " pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.389296 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt"] Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.389447 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0-apiservice-cert\") pod \"metallb-operator-webhook-server-685db896b9-sn7lp\" (UID: \"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0\") " pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.407827 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0-webhook-cert\") pod \"metallb-operator-webhook-server-685db896b9-sn7lp\" (UID: \"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0\") " pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.418239 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbp7f\" (UniqueName: \"kubernetes.io/projected/30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0-kube-api-access-fbp7f\") pod \"metallb-operator-webhook-server-685db896b9-sn7lp\" (UID: \"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0\") " pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.489676 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.534895 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" event={"ID":"ae4cf098-3915-418a-ae09-e1d08bfc523b","Type":"ContainerStarted","Data":"b1bacbd15018d57b1bdb4723ea05af40c7245a66b1a0dcd82d6c1542cdfb4967"} Dec 06 07:09:28 crc kubenswrapper[4945]: I1206 07:09:28.921921 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp"] Dec 06 07:09:28 crc kubenswrapper[4945]: W1206 07:09:28.927347 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30d3f7c2_7fc1_463c_9cf7_f7f6b2f74ab0.slice/crio-25bbf42874d06d535ad22e930e1d66f44009cef8b90d996b86adae682624aa41 WatchSource:0}: Error finding container 25bbf42874d06d535ad22e930e1d66f44009cef8b90d996b86adae682624aa41: Status 404 returned error can't find the container with id 25bbf42874d06d535ad22e930e1d66f44009cef8b90d996b86adae682624aa41 Dec 06 07:09:29 crc kubenswrapper[4945]: I1206 07:09:29.542096 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" event={"ID":"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0","Type":"ContainerStarted","Data":"25bbf42874d06d535ad22e930e1d66f44009cef8b90d996b86adae682624aa41"} Dec 06 07:09:32 crc kubenswrapper[4945]: I1206 07:09:32.562588 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" event={"ID":"ae4cf098-3915-418a-ae09-e1d08bfc523b","Type":"ContainerStarted","Data":"a74d73037c64ccedc4db39463588114e6a3194a81333617b36d1c284ad8c9b87"} Dec 06 07:09:32 crc kubenswrapper[4945]: I1206 07:09:32.562756 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:09:32 crc kubenswrapper[4945]: I1206 07:09:32.591971 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" podStartSLOduration=2.297798015 podStartE2EDuration="5.591948481s" podCreationTimestamp="2025-12-06 07:09:27 +0000 UTC" firstStartedPulling="2025-12-06 07:09:28.457473411 +0000 UTC m=+981.912334455" lastFinishedPulling="2025-12-06 07:09:31.751623877 +0000 UTC m=+985.206484921" observedRunningTime="2025-12-06 07:09:32.58596578 +0000 UTC m=+986.040826854" watchObservedRunningTime="2025-12-06 07:09:32.591948481 +0000 UTC m=+986.046809525" Dec 06 07:09:34 crc kubenswrapper[4945]: I1206 07:09:34.574638 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" event={"ID":"30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0","Type":"ContainerStarted","Data":"a9aef140b33b17540b3b8b8ecc0676853aeba6b7fa0625462f88cfd8db7646a2"} Dec 06 07:09:34 crc kubenswrapper[4945]: I1206 07:09:34.574994 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:09:34 crc kubenswrapper[4945]: I1206 07:09:34.597984 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" podStartSLOduration=1.290230882 podStartE2EDuration="6.597959479s" podCreationTimestamp="2025-12-06 07:09:28 +0000 UTC" firstStartedPulling="2025-12-06 07:09:28.930243918 +0000 UTC m=+982.385104952" lastFinishedPulling="2025-12-06 07:09:34.237972505 +0000 UTC m=+987.692833549" observedRunningTime="2025-12-06 07:09:34.591082073 +0000 UTC m=+988.045943117" watchObservedRunningTime="2025-12-06 07:09:34.597959479 +0000 UTC m=+988.052820523" Dec 06 07:09:38 crc kubenswrapper[4945]: I1206 07:09:38.795207 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:09:38 crc kubenswrapper[4945]: I1206 07:09:38.795787 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:09:48 crc kubenswrapper[4945]: I1206 07:09:48.497768 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-685db896b9-sn7lp" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.122327 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6dd5c5658d-plhqt" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.795885 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.796231 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.815977 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-269rv"] Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.818633 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.819830 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh"] Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.820594 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.823851 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.823858 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-ckq8q" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.823972 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.824005 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.833977 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh"] Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.911992 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-pfl8t"] Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.913196 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-pfl8t" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.915264 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.915546 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.915708 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-zzksh" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.916322 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.925152 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-m6c5w"] Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.926630 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.928309 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.933790 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2254f20-d815-4f62-98e5-3fb4bd20fd49-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6ljkh\" (UID: \"a2254f20-d815-4f62-98e5-3fb4bd20fd49\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.933857 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-metrics\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.933923 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-reloader\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.933975 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhftw\" (UniqueName: \"kubernetes.io/projected/58b9a31e-1915-40f7-8e45-cca0242826aa-kube-api-access-rhftw\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.934006 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-frr-conf\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.934122 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58b9a31e-1915-40f7-8e45-cca0242826aa-metrics-certs\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.934193 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gksmz\" (UniqueName: \"kubernetes.io/projected/a2254f20-d815-4f62-98e5-3fb4bd20fd49-kube-api-access-gksmz\") pod \"frr-k8s-webhook-server-7fcb986d4-6ljkh\" (UID: \"a2254f20-d815-4f62-98e5-3fb4bd20fd49\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.934226 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/58b9a31e-1915-40f7-8e45-cca0242826aa-frr-startup\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.934255 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-frr-sockets\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:08 crc kubenswrapper[4945]: I1206 07:10:08.947297 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-m6c5w"] Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035129 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-metrics-certs\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035200 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gksmz\" (UniqueName: \"kubernetes.io/projected/a2254f20-d815-4f62-98e5-3fb4bd20fd49-kube-api-access-gksmz\") pod \"frr-k8s-webhook-server-7fcb986d4-6ljkh\" (UID: \"a2254f20-d815-4f62-98e5-3fb4bd20fd49\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035223 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/58b9a31e-1915-40f7-8e45-cca0242826aa-frr-startup\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035245 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-frr-sockets\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035294 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98251a38-5e6f-4fff-b389-1ffec3864d22-cert\") pod \"controller-f8648f98b-m6c5w\" (UID: \"98251a38-5e6f-4fff-b389-1ffec3864d22\") " pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035324 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2254f20-d815-4f62-98e5-3fb4bd20fd49-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6ljkh\" (UID: \"a2254f20-d815-4f62-98e5-3fb4bd20fd49\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035348 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxrkw\" (UniqueName: \"kubernetes.io/projected/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-kube-api-access-qxrkw\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035394 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-metrics\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035418 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-metallb-excludel2\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035441 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-reloader\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: E1206 07:10:09.035457 4945 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 06 07:10:09 crc kubenswrapper[4945]: E1206 07:10:09.035539 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a2254f20-d815-4f62-98e5-3fb4bd20fd49-cert podName:a2254f20-d815-4f62-98e5-3fb4bd20fd49 nodeName:}" failed. No retries permitted until 2025-12-06 07:10:09.535519397 +0000 UTC m=+1022.990380441 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a2254f20-d815-4f62-98e5-3fb4bd20fd49-cert") pod "frr-k8s-webhook-server-7fcb986d4-6ljkh" (UID: "a2254f20-d815-4f62-98e5-3fb4bd20fd49") : secret "frr-k8s-webhook-server-cert" not found Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035469 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhftw\" (UniqueName: \"kubernetes.io/projected/58b9a31e-1915-40f7-8e45-cca0242826aa-kube-api-access-rhftw\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035723 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98251a38-5e6f-4fff-b389-1ffec3864d22-metrics-certs\") pod \"controller-f8648f98b-m6c5w\" (UID: \"98251a38-5e6f-4fff-b389-1ffec3864d22\") " pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035822 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-frr-conf\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035854 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58b9a31e-1915-40f7-8e45-cca0242826aa-metrics-certs\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035882 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxqkk\" (UniqueName: \"kubernetes.io/projected/98251a38-5e6f-4fff-b389-1ffec3864d22-kube-api-access-jxqkk\") pod \"controller-f8648f98b-m6c5w\" (UID: \"98251a38-5e6f-4fff-b389-1ffec3864d22\") " pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035932 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-memberlist\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.035952 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-frr-sockets\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.036027 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-metrics\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: E1206 07:10:09.036050 4945 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 06 07:10:09 crc kubenswrapper[4945]: E1206 07:10:09.036137 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58b9a31e-1915-40f7-8e45-cca0242826aa-metrics-certs podName:58b9a31e-1915-40f7-8e45-cca0242826aa nodeName:}" failed. No retries permitted until 2025-12-06 07:10:09.536108563 +0000 UTC m=+1022.990969677 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58b9a31e-1915-40f7-8e45-cca0242826aa-metrics-certs") pod "frr-k8s-269rv" (UID: "58b9a31e-1915-40f7-8e45-cca0242826aa") : secret "frr-k8s-certs-secret" not found Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.036181 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-frr-conf\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.036349 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/58b9a31e-1915-40f7-8e45-cca0242826aa-reloader\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.036638 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/58b9a31e-1915-40f7-8e45-cca0242826aa-frr-startup\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.059176 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhftw\" (UniqueName: \"kubernetes.io/projected/58b9a31e-1915-40f7-8e45-cca0242826aa-kube-api-access-rhftw\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.060810 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gksmz\" (UniqueName: \"kubernetes.io/projected/a2254f20-d815-4f62-98e5-3fb4bd20fd49-kube-api-access-gksmz\") pod \"frr-k8s-webhook-server-7fcb986d4-6ljkh\" (UID: \"a2254f20-d815-4f62-98e5-3fb4bd20fd49\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.137814 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98251a38-5e6f-4fff-b389-1ffec3864d22-cert\") pod \"controller-f8648f98b-m6c5w\" (UID: \"98251a38-5e6f-4fff-b389-1ffec3864d22\") " pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.137898 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxrkw\" (UniqueName: \"kubernetes.io/projected/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-kube-api-access-qxrkw\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.137939 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-metallb-excludel2\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.137975 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98251a38-5e6f-4fff-b389-1ffec3864d22-metrics-certs\") pod \"controller-f8648f98b-m6c5w\" (UID: \"98251a38-5e6f-4fff-b389-1ffec3864d22\") " pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.138035 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxqkk\" (UniqueName: \"kubernetes.io/projected/98251a38-5e6f-4fff-b389-1ffec3864d22-kube-api-access-jxqkk\") pod \"controller-f8648f98b-m6c5w\" (UID: \"98251a38-5e6f-4fff-b389-1ffec3864d22\") " pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.138078 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-memberlist\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.138105 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-metrics-certs\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: E1206 07:10:09.138416 4945 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 07:10:09 crc kubenswrapper[4945]: E1206 07:10:09.138469 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-memberlist podName:7e84c3f8-86bc-4ce1-9e48-936734a78b6d nodeName:}" failed. No retries permitted until 2025-12-06 07:10:09.638453795 +0000 UTC m=+1023.093314839 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-memberlist") pod "speaker-pfl8t" (UID: "7e84c3f8-86bc-4ce1-9e48-936734a78b6d") : secret "metallb-memberlist" not found Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.139355 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-metallb-excludel2\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.142667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-metrics-certs\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.142739 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98251a38-5e6f-4fff-b389-1ffec3864d22-metrics-certs\") pod \"controller-f8648f98b-m6c5w\" (UID: \"98251a38-5e6f-4fff-b389-1ffec3864d22\") " pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.142861 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.153031 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98251a38-5e6f-4fff-b389-1ffec3864d22-cert\") pod \"controller-f8648f98b-m6c5w\" (UID: \"98251a38-5e6f-4fff-b389-1ffec3864d22\") " pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.159185 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxrkw\" (UniqueName: \"kubernetes.io/projected/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-kube-api-access-qxrkw\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.161255 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxqkk\" (UniqueName: \"kubernetes.io/projected/98251a38-5e6f-4fff-b389-1ffec3864d22-kube-api-access-jxqkk\") pod \"controller-f8648f98b-m6c5w\" (UID: \"98251a38-5e6f-4fff-b389-1ffec3864d22\") " pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.244748 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.499089 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-m6c5w"] Dec 06 07:10:09 crc kubenswrapper[4945]: W1206 07:10:09.505324 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98251a38_5e6f_4fff_b389_1ffec3864d22.slice/crio-fd6263dedfc6688514bb0f213d2bc624d54f4a2dc0359362aaac1da60e62b0ec WatchSource:0}: Error finding container fd6263dedfc6688514bb0f213d2bc624d54f4a2dc0359362aaac1da60e62b0ec: Status 404 returned error can't find the container with id fd6263dedfc6688514bb0f213d2bc624d54f4a2dc0359362aaac1da60e62b0ec Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.542369 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58b9a31e-1915-40f7-8e45-cca0242826aa-metrics-certs\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.542553 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2254f20-d815-4f62-98e5-3fb4bd20fd49-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6ljkh\" (UID: \"a2254f20-d815-4f62-98e5-3fb4bd20fd49\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.549889 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58b9a31e-1915-40f7-8e45-cca0242826aa-metrics-certs\") pod \"frr-k8s-269rv\" (UID: \"58b9a31e-1915-40f7-8e45-cca0242826aa\") " pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.553019 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2254f20-d815-4f62-98e5-3fb4bd20fd49-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6ljkh\" (UID: \"a2254f20-d815-4f62-98e5-3fb4bd20fd49\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.644334 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-memberlist\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:09 crc kubenswrapper[4945]: E1206 07:10:09.644505 4945 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 07:10:09 crc kubenswrapper[4945]: E1206 07:10:09.644582 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-memberlist podName:7e84c3f8-86bc-4ce1-9e48-936734a78b6d nodeName:}" failed. No retries permitted until 2025-12-06 07:10:10.644565551 +0000 UTC m=+1024.099426595 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-memberlist") pod "speaker-pfl8t" (UID: "7e84c3f8-86bc-4ce1-9e48-936734a78b6d") : secret "metallb-memberlist" not found Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.740315 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.750223 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:09 crc kubenswrapper[4945]: I1206 07:10:09.794408 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-m6c5w" event={"ID":"98251a38-5e6f-4fff-b389-1ffec3864d22","Type":"ContainerStarted","Data":"fd6263dedfc6688514bb0f213d2bc624d54f4a2dc0359362aaac1da60e62b0ec"} Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.062864 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh"] Dec 06 07:10:10 crc kubenswrapper[4945]: W1206 07:10:10.072397 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2254f20_d815_4f62_98e5_3fb4bd20fd49.slice/crio-7deb22ad12a05ef63e2c9aa834a1999e6f20fdc4e3dda7e1bd4b632a87ca6c80 WatchSource:0}: Error finding container 7deb22ad12a05ef63e2c9aa834a1999e6f20fdc4e3dda7e1bd4b632a87ca6c80: Status 404 returned error can't find the container with id 7deb22ad12a05ef63e2c9aa834a1999e6f20fdc4e3dda7e1bd4b632a87ca6c80 Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.658514 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-memberlist\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.664429 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7e84c3f8-86bc-4ce1-9e48-936734a78b6d-memberlist\") pod \"speaker-pfl8t\" (UID: \"7e84c3f8-86bc-4ce1-9e48-936734a78b6d\") " pod="metallb-system/speaker-pfl8t" Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.729851 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-pfl8t" Dec 06 07:10:10 crc kubenswrapper[4945]: W1206 07:10:10.752802 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e84c3f8_86bc_4ce1_9e48_936734a78b6d.slice/crio-0cb39695f5a15531abde9c3edf2fe1e48dfcfa8deb432f5668223a2032cdf4e0 WatchSource:0}: Error finding container 0cb39695f5a15531abde9c3edf2fe1e48dfcfa8deb432f5668223a2032cdf4e0: Status 404 returned error can't find the container with id 0cb39695f5a15531abde9c3edf2fe1e48dfcfa8deb432f5668223a2032cdf4e0 Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.802091 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-pfl8t" event={"ID":"7e84c3f8-86bc-4ce1-9e48-936734a78b6d","Type":"ContainerStarted","Data":"0cb39695f5a15531abde9c3edf2fe1e48dfcfa8deb432f5668223a2032cdf4e0"} Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.803870 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" event={"ID":"a2254f20-d815-4f62-98e5-3fb4bd20fd49","Type":"ContainerStarted","Data":"7deb22ad12a05ef63e2c9aa834a1999e6f20fdc4e3dda7e1bd4b632a87ca6c80"} Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.805181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-269rv" event={"ID":"58b9a31e-1915-40f7-8e45-cca0242826aa","Type":"ContainerStarted","Data":"c6f1d89f00fe7085947264ab4e81e464e9a504072a5e3db2192a72cd79dd7439"} Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.810012 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-m6c5w" event={"ID":"98251a38-5e6f-4fff-b389-1ffec3864d22","Type":"ContainerStarted","Data":"e304a8e093d26b44b50a4b889232e91a03dfcc6683e7382f03ffde238323fcb4"} Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.810055 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-m6c5w" event={"ID":"98251a38-5e6f-4fff-b389-1ffec3864d22","Type":"ContainerStarted","Data":"7d9dfc60f3c6faaaeafa71a6201d35f1c26605e3baa2cca7f75acfc7ebf82706"} Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.810145 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:10 crc kubenswrapper[4945]: I1206 07:10:10.829247 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-m6c5w" podStartSLOduration=2.829208766 podStartE2EDuration="2.829208766s" podCreationTimestamp="2025-12-06 07:10:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:10:10.828914208 +0000 UTC m=+1024.283775272" watchObservedRunningTime="2025-12-06 07:10:10.829208766 +0000 UTC m=+1024.284069810" Dec 06 07:10:11 crc kubenswrapper[4945]: I1206 07:10:11.821936 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-pfl8t" event={"ID":"7e84c3f8-86bc-4ce1-9e48-936734a78b6d","Type":"ContainerStarted","Data":"5b2b4b7c34872b365edf21b8af8168259cf2916f4dd26cd827f7d51965b84c4e"} Dec 06 07:10:11 crc kubenswrapper[4945]: I1206 07:10:11.821989 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-pfl8t" event={"ID":"7e84c3f8-86bc-4ce1-9e48-936734a78b6d","Type":"ContainerStarted","Data":"a873acece89e1145274bf2c052fdd313515a3a1e78c60de7515bccebf5386544"} Dec 06 07:10:11 crc kubenswrapper[4945]: I1206 07:10:11.822020 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-pfl8t" Dec 06 07:10:11 crc kubenswrapper[4945]: I1206 07:10:11.865569 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-pfl8t" podStartSLOduration=3.865548429 podStartE2EDuration="3.865548429s" podCreationTimestamp="2025-12-06 07:10:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:10:11.861750277 +0000 UTC m=+1025.316611321" watchObservedRunningTime="2025-12-06 07:10:11.865548429 +0000 UTC m=+1025.320409473" Dec 06 07:10:19 crc kubenswrapper[4945]: I1206 07:10:19.248876 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-m6c5w" Dec 06 07:10:20 crc kubenswrapper[4945]: I1206 07:10:20.733858 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-pfl8t" Dec 06 07:10:21 crc kubenswrapper[4945]: I1206 07:10:21.925382 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" event={"ID":"a2254f20-d815-4f62-98e5-3fb4bd20fd49","Type":"ContainerStarted","Data":"60264cf33fcf9716dfc56c9e4b1669ad2443397996b104ebecfa4b3178e63073"} Dec 06 07:10:21 crc kubenswrapper[4945]: I1206 07:10:21.925799 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:21 crc kubenswrapper[4945]: I1206 07:10:21.927497 4945 generic.go:334] "Generic (PLEG): container finished" podID="58b9a31e-1915-40f7-8e45-cca0242826aa" containerID="c6adb3f5ec0d6d0a41b4a990e64ee897c6fcd900089cb357cf04b7f4432d7548" exitCode=0 Dec 06 07:10:21 crc kubenswrapper[4945]: I1206 07:10:21.927525 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-269rv" event={"ID":"58b9a31e-1915-40f7-8e45-cca0242826aa","Type":"ContainerDied","Data":"c6adb3f5ec0d6d0a41b4a990e64ee897c6fcd900089cb357cf04b7f4432d7548"} Dec 06 07:10:21 crc kubenswrapper[4945]: I1206 07:10:21.945147 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" podStartSLOduration=2.952674017 podStartE2EDuration="13.945116463s" podCreationTimestamp="2025-12-06 07:10:08 +0000 UTC" firstStartedPulling="2025-12-06 07:10:10.074985615 +0000 UTC m=+1023.529846659" lastFinishedPulling="2025-12-06 07:10:21.067428061 +0000 UTC m=+1034.522289105" observedRunningTime="2025-12-06 07:10:21.940910089 +0000 UTC m=+1035.395771133" watchObservedRunningTime="2025-12-06 07:10:21.945116463 +0000 UTC m=+1035.399977507" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.337208 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44"] Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.338743 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.341451 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.350592 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44"] Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.468756 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.468860 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.468923 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg2zd\" (UniqueName: \"kubernetes.io/projected/0549fa78-72df-422e-b112-cdff8788ebcb-kube-api-access-dg2zd\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.570010 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.570090 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg2zd\" (UniqueName: \"kubernetes.io/projected/0549fa78-72df-422e-b112-cdff8788ebcb-kube-api-access-dg2zd\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.570170 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.570652 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.570655 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.591088 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg2zd\" (UniqueName: \"kubernetes.io/projected/0549fa78-72df-422e-b112-cdff8788ebcb-kube-api-access-dg2zd\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.667563 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.870421 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44"] Dec 06 07:10:22 crc kubenswrapper[4945]: W1206 07:10:22.872864 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0549fa78_72df_422e_b112_cdff8788ebcb.slice/crio-98c0e9e00b6132abf20064b813f49fc4df72dccb15d042d23f4e280a4e3bd653 WatchSource:0}: Error finding container 98c0e9e00b6132abf20064b813f49fc4df72dccb15d042d23f4e280a4e3bd653: Status 404 returned error can't find the container with id 98c0e9e00b6132abf20064b813f49fc4df72dccb15d042d23f4e280a4e3bd653 Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.935209 4945 generic.go:334] "Generic (PLEG): container finished" podID="58b9a31e-1915-40f7-8e45-cca0242826aa" containerID="65fd1d15f223a7b0257c485768d36aad8871ec191b5cd1e0bdf88c4590cbf81f" exitCode=0 Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.935306 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-269rv" event={"ID":"58b9a31e-1915-40f7-8e45-cca0242826aa","Type":"ContainerDied","Data":"65fd1d15f223a7b0257c485768d36aad8871ec191b5cd1e0bdf88c4590cbf81f"} Dec 06 07:10:22 crc kubenswrapper[4945]: I1206 07:10:22.937911 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" event={"ID":"0549fa78-72df-422e-b112-cdff8788ebcb","Type":"ContainerStarted","Data":"98c0e9e00b6132abf20064b813f49fc4df72dccb15d042d23f4e280a4e3bd653"} Dec 06 07:10:23 crc kubenswrapper[4945]: I1206 07:10:23.945905 4945 generic.go:334] "Generic (PLEG): container finished" podID="58b9a31e-1915-40f7-8e45-cca0242826aa" containerID="dc6ed1805ccb927761c63cc3f5111e6ac92eb6d888801300c298208e4eb7d889" exitCode=0 Dec 06 07:10:23 crc kubenswrapper[4945]: I1206 07:10:23.945982 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-269rv" event={"ID":"58b9a31e-1915-40f7-8e45-cca0242826aa","Type":"ContainerDied","Data":"dc6ed1805ccb927761c63cc3f5111e6ac92eb6d888801300c298208e4eb7d889"} Dec 06 07:10:23 crc kubenswrapper[4945]: I1206 07:10:23.948682 4945 generic.go:334] "Generic (PLEG): container finished" podID="0549fa78-72df-422e-b112-cdff8788ebcb" containerID="c7c40cf2dcd59b697e1a3472cd3975e44b5cd5276e500f8cef6542e3f976f943" exitCode=0 Dec 06 07:10:23 crc kubenswrapper[4945]: I1206 07:10:23.948729 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" event={"ID":"0549fa78-72df-422e-b112-cdff8788ebcb","Type":"ContainerDied","Data":"c7c40cf2dcd59b697e1a3472cd3975e44b5cd5276e500f8cef6542e3f976f943"} Dec 06 07:10:24 crc kubenswrapper[4945]: I1206 07:10:24.961242 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-269rv" event={"ID":"58b9a31e-1915-40f7-8e45-cca0242826aa","Type":"ContainerStarted","Data":"c444c2afb80e07c37426ba7fe2b2e1301c129af94b8eb455d8dd224cca67f387"} Dec 06 07:10:24 crc kubenswrapper[4945]: I1206 07:10:24.962346 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-269rv" event={"ID":"58b9a31e-1915-40f7-8e45-cca0242826aa","Type":"ContainerStarted","Data":"539c0757f3c5d54e3989dc3f31b833fce6ae3fd1cc89c24c92dff4a823165263"} Dec 06 07:10:24 crc kubenswrapper[4945]: I1206 07:10:24.962414 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-269rv" event={"ID":"58b9a31e-1915-40f7-8e45-cca0242826aa","Type":"ContainerStarted","Data":"081b6ef5a72dbb786838de958c893e9bb96e44e4a212376c98f25e816b7f4fd0"} Dec 06 07:10:24 crc kubenswrapper[4945]: I1206 07:10:24.962423 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-269rv" event={"ID":"58b9a31e-1915-40f7-8e45-cca0242826aa","Type":"ContainerStarted","Data":"b3738610c0f99896a0128b5f32e4b9c294b166ecbde7abbdc404e2269ed0b367"} Dec 06 07:10:25 crc kubenswrapper[4945]: I1206 07:10:25.975746 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-269rv" event={"ID":"58b9a31e-1915-40f7-8e45-cca0242826aa","Type":"ContainerStarted","Data":"46708e7103b18ba2c3cf1a6d3b4e19dddfaa15b3bd1e8457889a75efccaa5fe0"} Dec 06 07:10:25 crc kubenswrapper[4945]: I1206 07:10:25.976338 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-269rv" event={"ID":"58b9a31e-1915-40f7-8e45-cca0242826aa","Type":"ContainerStarted","Data":"f1ac4d124127991442dfbf88c96687f26a12dbe7b29742147cc9ea9e2f3df91a"} Dec 06 07:10:26 crc kubenswrapper[4945]: I1206 07:10:26.004645 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-269rv" podStartSLOduration=6.887001475 podStartE2EDuration="18.004611928s" podCreationTimestamp="2025-12-06 07:10:08 +0000 UTC" firstStartedPulling="2025-12-06 07:10:09.970502546 +0000 UTC m=+1023.425363580" lastFinishedPulling="2025-12-06 07:10:21.088112989 +0000 UTC m=+1034.542974033" observedRunningTime="2025-12-06 07:10:26.003950021 +0000 UTC m=+1039.458811075" watchObservedRunningTime="2025-12-06 07:10:26.004611928 +0000 UTC m=+1039.459472982" Dec 06 07:10:26 crc kubenswrapper[4945]: I1206 07:10:26.981977 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:29 crc kubenswrapper[4945]: I1206 07:10:29.007442 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" event={"ID":"0549fa78-72df-422e-b112-cdff8788ebcb","Type":"ContainerStarted","Data":"2dd830a8b98fda74b259f44d5cf9a1b0678f8dbba45e9b52fe3a8a6616eed58e"} Dec 06 07:10:29 crc kubenswrapper[4945]: I1206 07:10:29.741439 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:29 crc kubenswrapper[4945]: I1206 07:10:29.796919 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:30 crc kubenswrapper[4945]: I1206 07:10:30.014976 4945 generic.go:334] "Generic (PLEG): container finished" podID="0549fa78-72df-422e-b112-cdff8788ebcb" containerID="2dd830a8b98fda74b259f44d5cf9a1b0678f8dbba45e9b52fe3a8a6616eed58e" exitCode=0 Dec 06 07:10:30 crc kubenswrapper[4945]: I1206 07:10:30.015073 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" event={"ID":"0549fa78-72df-422e-b112-cdff8788ebcb","Type":"ContainerDied","Data":"2dd830a8b98fda74b259f44d5cf9a1b0678f8dbba45e9b52fe3a8a6616eed58e"} Dec 06 07:10:31 crc kubenswrapper[4945]: I1206 07:10:31.024220 4945 generic.go:334] "Generic (PLEG): container finished" podID="0549fa78-72df-422e-b112-cdff8788ebcb" containerID="daa13f162d5b8fd7b5eb9b2ff2499740d325b2c534eed53b383ac4e44e5837a3" exitCode=0 Dec 06 07:10:31 crc kubenswrapper[4945]: I1206 07:10:31.024269 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" event={"ID":"0549fa78-72df-422e-b112-cdff8788ebcb","Type":"ContainerDied","Data":"daa13f162d5b8fd7b5eb9b2ff2499740d325b2c534eed53b383ac4e44e5837a3"} Dec 06 07:10:32 crc kubenswrapper[4945]: I1206 07:10:32.261568 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:32 crc kubenswrapper[4945]: I1206 07:10:32.411817 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-bundle\") pod \"0549fa78-72df-422e-b112-cdff8788ebcb\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " Dec 06 07:10:32 crc kubenswrapper[4945]: I1206 07:10:32.412108 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg2zd\" (UniqueName: \"kubernetes.io/projected/0549fa78-72df-422e-b112-cdff8788ebcb-kube-api-access-dg2zd\") pod \"0549fa78-72df-422e-b112-cdff8788ebcb\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " Dec 06 07:10:32 crc kubenswrapper[4945]: I1206 07:10:32.412289 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-util\") pod \"0549fa78-72df-422e-b112-cdff8788ebcb\" (UID: \"0549fa78-72df-422e-b112-cdff8788ebcb\") " Dec 06 07:10:32 crc kubenswrapper[4945]: I1206 07:10:32.412832 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-bundle" (OuterVolumeSpecName: "bundle") pod "0549fa78-72df-422e-b112-cdff8788ebcb" (UID: "0549fa78-72df-422e-b112-cdff8788ebcb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:10:32 crc kubenswrapper[4945]: I1206 07:10:32.417479 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0549fa78-72df-422e-b112-cdff8788ebcb-kube-api-access-dg2zd" (OuterVolumeSpecName: "kube-api-access-dg2zd") pod "0549fa78-72df-422e-b112-cdff8788ebcb" (UID: "0549fa78-72df-422e-b112-cdff8788ebcb"). InnerVolumeSpecName "kube-api-access-dg2zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:10:32 crc kubenswrapper[4945]: I1206 07:10:32.424144 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-util" (OuterVolumeSpecName: "util") pod "0549fa78-72df-422e-b112-cdff8788ebcb" (UID: "0549fa78-72df-422e-b112-cdff8788ebcb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:10:32 crc kubenswrapper[4945]: I1206 07:10:32.513499 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:10:32 crc kubenswrapper[4945]: I1206 07:10:32.513548 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg2zd\" (UniqueName: \"kubernetes.io/projected/0549fa78-72df-422e-b112-cdff8788ebcb-kube-api-access-dg2zd\") on node \"crc\" DevicePath \"\"" Dec 06 07:10:32 crc kubenswrapper[4945]: I1206 07:10:32.513564 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0549fa78-72df-422e-b112-cdff8788ebcb-util\") on node \"crc\" DevicePath \"\"" Dec 06 07:10:33 crc kubenswrapper[4945]: I1206 07:10:33.039006 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" event={"ID":"0549fa78-72df-422e-b112-cdff8788ebcb","Type":"ContainerDied","Data":"98c0e9e00b6132abf20064b813f49fc4df72dccb15d042d23f4e280a4e3bd653"} Dec 06 07:10:33 crc kubenswrapper[4945]: I1206 07:10:33.039052 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98c0e9e00b6132abf20064b813f49fc4df72dccb15d042d23f4e280a4e3bd653" Dec 06 07:10:33 crc kubenswrapper[4945]: I1206 07:10:33.039402 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44" Dec 06 07:10:38 crc kubenswrapper[4945]: I1206 07:10:38.796058 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:10:38 crc kubenswrapper[4945]: I1206 07:10:38.796646 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:10:38 crc kubenswrapper[4945]: I1206 07:10:38.796691 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:10:38 crc kubenswrapper[4945]: I1206 07:10:38.797249 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fcc05253d8d77edb32c43d3ca8c759b1e83d7e91658c6a423615205b104949ce"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:10:38 crc kubenswrapper[4945]: I1206 07:10:38.797320 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://fcc05253d8d77edb32c43d3ca8c759b1e83d7e91658c6a423615205b104949ce" gracePeriod=600 Dec 06 07:10:39 crc kubenswrapper[4945]: I1206 07:10:39.075896 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="fcc05253d8d77edb32c43d3ca8c759b1e83d7e91658c6a423615205b104949ce" exitCode=0 Dec 06 07:10:39 crc kubenswrapper[4945]: I1206 07:10:39.075928 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"fcc05253d8d77edb32c43d3ca8c759b1e83d7e91658c6a423615205b104949ce"} Dec 06 07:10:39 crc kubenswrapper[4945]: I1206 07:10:39.076297 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"fd69474a2e644df9791d0d3744c09cd9214dcea77b7b07f5d390844d87230bc1"} Dec 06 07:10:39 crc kubenswrapper[4945]: I1206 07:10:39.076329 4945 scope.go:117] "RemoveContainer" containerID="0c1b134b46f37e7ead92dfbee219cc6a32bd1e1e7b37766e6910364e7c42deb2" Dec 06 07:10:39 crc kubenswrapper[4945]: I1206 07:10:39.743336 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-269rv" Dec 06 07:10:39 crc kubenswrapper[4945]: I1206 07:10:39.754836 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6ljkh" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.870438 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm"] Dec 06 07:10:40 crc kubenswrapper[4945]: E1206 07:10:40.870695 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0549fa78-72df-422e-b112-cdff8788ebcb" containerName="pull" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.870708 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0549fa78-72df-422e-b112-cdff8788ebcb" containerName="pull" Dec 06 07:10:40 crc kubenswrapper[4945]: E1206 07:10:40.870720 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0549fa78-72df-422e-b112-cdff8788ebcb" containerName="extract" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.870726 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0549fa78-72df-422e-b112-cdff8788ebcb" containerName="extract" Dec 06 07:10:40 crc kubenswrapper[4945]: E1206 07:10:40.870741 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0549fa78-72df-422e-b112-cdff8788ebcb" containerName="util" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.870747 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0549fa78-72df-422e-b112-cdff8788ebcb" containerName="util" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.870856 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0549fa78-72df-422e-b112-cdff8788ebcb" containerName="extract" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.871302 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.873978 4945 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-8st2q" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.874664 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.875350 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.918592 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/89eebed1-9dfd-48c0-9caf-c51218f3e5ad-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-hl8cm\" (UID: \"89eebed1-9dfd-48c0-9caf-c51218f3e5ad\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.918936 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6jnb\" (UniqueName: \"kubernetes.io/projected/89eebed1-9dfd-48c0-9caf-c51218f3e5ad-kube-api-access-p6jnb\") pod \"cert-manager-operator-controller-manager-64cf6dff88-hl8cm\" (UID: \"89eebed1-9dfd-48c0-9caf-c51218f3e5ad\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" Dec 06 07:10:40 crc kubenswrapper[4945]: I1206 07:10:40.937909 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm"] Dec 06 07:10:41 crc kubenswrapper[4945]: I1206 07:10:41.020479 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6jnb\" (UniqueName: \"kubernetes.io/projected/89eebed1-9dfd-48c0-9caf-c51218f3e5ad-kube-api-access-p6jnb\") pod \"cert-manager-operator-controller-manager-64cf6dff88-hl8cm\" (UID: \"89eebed1-9dfd-48c0-9caf-c51218f3e5ad\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" Dec 06 07:10:41 crc kubenswrapper[4945]: I1206 07:10:41.020617 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/89eebed1-9dfd-48c0-9caf-c51218f3e5ad-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-hl8cm\" (UID: \"89eebed1-9dfd-48c0-9caf-c51218f3e5ad\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" Dec 06 07:10:41 crc kubenswrapper[4945]: I1206 07:10:41.022554 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/89eebed1-9dfd-48c0-9caf-c51218f3e5ad-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-hl8cm\" (UID: \"89eebed1-9dfd-48c0-9caf-c51218f3e5ad\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" Dec 06 07:10:41 crc kubenswrapper[4945]: I1206 07:10:41.045258 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6jnb\" (UniqueName: \"kubernetes.io/projected/89eebed1-9dfd-48c0-9caf-c51218f3e5ad-kube-api-access-p6jnb\") pod \"cert-manager-operator-controller-manager-64cf6dff88-hl8cm\" (UID: \"89eebed1-9dfd-48c0-9caf-c51218f3e5ad\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" Dec 06 07:10:41 crc kubenswrapper[4945]: I1206 07:10:41.193317 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" Dec 06 07:10:41 crc kubenswrapper[4945]: I1206 07:10:41.620654 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm"] Dec 06 07:10:41 crc kubenswrapper[4945]: W1206 07:10:41.626552 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89eebed1_9dfd_48c0_9caf_c51218f3e5ad.slice/crio-0d4bfba63300730f16c213254ac6baa0018596239477085f9bc74e89ccf91a44 WatchSource:0}: Error finding container 0d4bfba63300730f16c213254ac6baa0018596239477085f9bc74e89ccf91a44: Status 404 returned error can't find the container with id 0d4bfba63300730f16c213254ac6baa0018596239477085f9bc74e89ccf91a44 Dec 06 07:10:42 crc kubenswrapper[4945]: I1206 07:10:42.098096 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" event={"ID":"89eebed1-9dfd-48c0-9caf-c51218f3e5ad","Type":"ContainerStarted","Data":"0d4bfba63300730f16c213254ac6baa0018596239477085f9bc74e89ccf91a44"} Dec 06 07:10:52 crc kubenswrapper[4945]: I1206 07:10:52.160579 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" event={"ID":"89eebed1-9dfd-48c0-9caf-c51218f3e5ad","Type":"ContainerStarted","Data":"d97fcdb95487f0dc58c6aa1a31df3b28e7dbb3442f630cd11a182797856d3392"} Dec 06 07:10:52 crc kubenswrapper[4945]: I1206 07:10:52.188592 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-hl8cm" podStartSLOduration=2.648471254 podStartE2EDuration="12.188568722s" podCreationTimestamp="2025-12-06 07:10:40 +0000 UTC" firstStartedPulling="2025-12-06 07:10:41.62953486 +0000 UTC m=+1055.084395904" lastFinishedPulling="2025-12-06 07:10:51.169632328 +0000 UTC m=+1064.624493372" observedRunningTime="2025-12-06 07:10:52.182965961 +0000 UTC m=+1065.637827005" watchObservedRunningTime="2025-12-06 07:10:52.188568722 +0000 UTC m=+1065.643429766" Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.682418 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-ms6cl"] Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.683752 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.690389 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.690531 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.690559 4945 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-hnwzn" Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.707760 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-ms6cl"] Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.811437 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96f7adf5-11a3-480d-ae16-cb4bf53ab693-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-ms6cl\" (UID: \"96f7adf5-11a3-480d-ae16-cb4bf53ab693\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.811537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2pk5\" (UniqueName: \"kubernetes.io/projected/96f7adf5-11a3-480d-ae16-cb4bf53ab693-kube-api-access-t2pk5\") pod \"cert-manager-webhook-f4fb5df64-ms6cl\" (UID: \"96f7adf5-11a3-480d-ae16-cb4bf53ab693\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.912455 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2pk5\" (UniqueName: \"kubernetes.io/projected/96f7adf5-11a3-480d-ae16-cb4bf53ab693-kube-api-access-t2pk5\") pod \"cert-manager-webhook-f4fb5df64-ms6cl\" (UID: \"96f7adf5-11a3-480d-ae16-cb4bf53ab693\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.912586 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96f7adf5-11a3-480d-ae16-cb4bf53ab693-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-ms6cl\" (UID: \"96f7adf5-11a3-480d-ae16-cb4bf53ab693\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.930817 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96f7adf5-11a3-480d-ae16-cb4bf53ab693-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-ms6cl\" (UID: \"96f7adf5-11a3-480d-ae16-cb4bf53ab693\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" Dec 06 07:10:54 crc kubenswrapper[4945]: I1206 07:10:54.931095 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2pk5\" (UniqueName: \"kubernetes.io/projected/96f7adf5-11a3-480d-ae16-cb4bf53ab693-kube-api-access-t2pk5\") pod \"cert-manager-webhook-f4fb5df64-ms6cl\" (UID: \"96f7adf5-11a3-480d-ae16-cb4bf53ab693\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" Dec 06 07:10:55 crc kubenswrapper[4945]: I1206 07:10:55.003100 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" Dec 06 07:10:55 crc kubenswrapper[4945]: I1206 07:10:55.795236 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-ms6cl"] Dec 06 07:10:55 crc kubenswrapper[4945]: W1206 07:10:55.816270 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96f7adf5_11a3_480d_ae16_cb4bf53ab693.slice/crio-a2f490dfa7a218f61365f2e0eb5f48f77288482ca2289a9596fd3f2831440812 WatchSource:0}: Error finding container a2f490dfa7a218f61365f2e0eb5f48f77288482ca2289a9596fd3f2831440812: Status 404 returned error can't find the container with id a2f490dfa7a218f61365f2e0eb5f48f77288482ca2289a9596fd3f2831440812 Dec 06 07:10:56 crc kubenswrapper[4945]: I1206 07:10:56.186742 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" event={"ID":"96f7adf5-11a3-480d-ae16-cb4bf53ab693","Type":"ContainerStarted","Data":"a2f490dfa7a218f61365f2e0eb5f48f77288482ca2289a9596fd3f2831440812"} Dec 06 07:10:57 crc kubenswrapper[4945]: I1206 07:10:57.877215 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt"] Dec 06 07:10:57 crc kubenswrapper[4945]: I1206 07:10:57.879161 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" Dec 06 07:10:57 crc kubenswrapper[4945]: I1206 07:10:57.880869 4945 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-52ncw" Dec 06 07:10:57 crc kubenswrapper[4945]: I1206 07:10:57.892572 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt"] Dec 06 07:10:57 crc kubenswrapper[4945]: I1206 07:10:57.992874 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a40739-7cbd-4492-aff7-ab1dd5bd63c8-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-qh9wt\" (UID: \"02a40739-7cbd-4492-aff7-ab1dd5bd63c8\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" Dec 06 07:10:57 crc kubenswrapper[4945]: I1206 07:10:57.993031 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k69ds\" (UniqueName: \"kubernetes.io/projected/02a40739-7cbd-4492-aff7-ab1dd5bd63c8-kube-api-access-k69ds\") pod \"cert-manager-cainjector-855d9ccff4-qh9wt\" (UID: \"02a40739-7cbd-4492-aff7-ab1dd5bd63c8\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" Dec 06 07:10:58 crc kubenswrapper[4945]: I1206 07:10:58.095795 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k69ds\" (UniqueName: \"kubernetes.io/projected/02a40739-7cbd-4492-aff7-ab1dd5bd63c8-kube-api-access-k69ds\") pod \"cert-manager-cainjector-855d9ccff4-qh9wt\" (UID: \"02a40739-7cbd-4492-aff7-ab1dd5bd63c8\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" Dec 06 07:10:58 crc kubenswrapper[4945]: I1206 07:10:58.095909 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a40739-7cbd-4492-aff7-ab1dd5bd63c8-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-qh9wt\" (UID: \"02a40739-7cbd-4492-aff7-ab1dd5bd63c8\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" Dec 06 07:10:58 crc kubenswrapper[4945]: I1206 07:10:58.124099 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a40739-7cbd-4492-aff7-ab1dd5bd63c8-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-qh9wt\" (UID: \"02a40739-7cbd-4492-aff7-ab1dd5bd63c8\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" Dec 06 07:10:58 crc kubenswrapper[4945]: I1206 07:10:58.278794 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k69ds\" (UniqueName: \"kubernetes.io/projected/02a40739-7cbd-4492-aff7-ab1dd5bd63c8-kube-api-access-k69ds\") pod \"cert-manager-cainjector-855d9ccff4-qh9wt\" (UID: \"02a40739-7cbd-4492-aff7-ab1dd5bd63c8\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" Dec 06 07:10:58 crc kubenswrapper[4945]: I1206 07:10:58.505190 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" Dec 06 07:10:59 crc kubenswrapper[4945]: I1206 07:10:59.100065 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt"] Dec 06 07:10:59 crc kubenswrapper[4945]: W1206 07:10:59.118336 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02a40739_7cbd_4492_aff7_ab1dd5bd63c8.slice/crio-c86f0b3546ae20051d0823c0744968e87277de529c7228468baaa749a336187b WatchSource:0}: Error finding container c86f0b3546ae20051d0823c0744968e87277de529c7228468baaa749a336187b: Status 404 returned error can't find the container with id c86f0b3546ae20051d0823c0744968e87277de529c7228468baaa749a336187b Dec 06 07:10:59 crc kubenswrapper[4945]: I1206 07:10:59.293618 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" event={"ID":"02a40739-7cbd-4492-aff7-ab1dd5bd63c8","Type":"ContainerStarted","Data":"c86f0b3546ae20051d0823c0744968e87277de529c7228468baaa749a336187b"} Dec 06 07:11:10 crc kubenswrapper[4945]: I1206 07:11:10.364148 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" event={"ID":"96f7adf5-11a3-480d-ae16-cb4bf53ab693","Type":"ContainerStarted","Data":"db224f5f1388e89603bab819a8529529f769fcd29264f21f7bf0a7f7b9fd0f0a"} Dec 06 07:11:10 crc kubenswrapper[4945]: I1206 07:11:10.364891 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" Dec 06 07:11:10 crc kubenswrapper[4945]: I1206 07:11:10.368226 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" event={"ID":"02a40739-7cbd-4492-aff7-ab1dd5bd63c8","Type":"ContainerStarted","Data":"68631fa1f4170d2ba8273a48a87aff90b9d88fa27e4b9e5a387a6d7316586758"} Dec 06 07:11:10 crc kubenswrapper[4945]: I1206 07:11:10.386035 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" podStartSLOduration=3.099916393 podStartE2EDuration="16.386015537s" podCreationTimestamp="2025-12-06 07:10:54 +0000 UTC" firstStartedPulling="2025-12-06 07:10:55.824635953 +0000 UTC m=+1069.279496997" lastFinishedPulling="2025-12-06 07:11:09.110735087 +0000 UTC m=+1082.565596141" observedRunningTime="2025-12-06 07:11:10.383385236 +0000 UTC m=+1083.838246290" watchObservedRunningTime="2025-12-06 07:11:10.386015537 +0000 UTC m=+1083.840876581" Dec 06 07:11:10 crc kubenswrapper[4945]: I1206 07:11:10.407648 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-qh9wt" podStartSLOduration=3.43864667 podStartE2EDuration="13.40762134s" podCreationTimestamp="2025-12-06 07:10:57 +0000 UTC" firstStartedPulling="2025-12-06 07:10:59.121928802 +0000 UTC m=+1072.576789846" lastFinishedPulling="2025-12-06 07:11:09.090903472 +0000 UTC m=+1082.545764516" observedRunningTime="2025-12-06 07:11:10.402910423 +0000 UTC m=+1083.857771467" watchObservedRunningTime="2025-12-06 07:11:10.40762134 +0000 UTC m=+1083.862482384" Dec 06 07:11:13 crc kubenswrapper[4945]: I1206 07:11:13.832847 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-cpjql"] Dec 06 07:11:13 crc kubenswrapper[4945]: I1206 07:11:13.833995 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-cpjql" Dec 06 07:11:13 crc kubenswrapper[4945]: I1206 07:11:13.836608 4945 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-v24qg" Dec 06 07:11:13 crc kubenswrapper[4945]: I1206 07:11:13.844661 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-cpjql"] Dec 06 07:11:13 crc kubenswrapper[4945]: I1206 07:11:13.980716 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q87sc\" (UniqueName: \"kubernetes.io/projected/8ca8d146-bf33-4f3e-94e2-a794187ee78e-kube-api-access-q87sc\") pod \"cert-manager-86cb77c54b-cpjql\" (UID: \"8ca8d146-bf33-4f3e-94e2-a794187ee78e\") " pod="cert-manager/cert-manager-86cb77c54b-cpjql" Dec 06 07:11:13 crc kubenswrapper[4945]: I1206 07:11:13.981303 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8ca8d146-bf33-4f3e-94e2-a794187ee78e-bound-sa-token\") pod \"cert-manager-86cb77c54b-cpjql\" (UID: \"8ca8d146-bf33-4f3e-94e2-a794187ee78e\") " pod="cert-manager/cert-manager-86cb77c54b-cpjql" Dec 06 07:11:14 crc kubenswrapper[4945]: I1206 07:11:14.082636 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q87sc\" (UniqueName: \"kubernetes.io/projected/8ca8d146-bf33-4f3e-94e2-a794187ee78e-kube-api-access-q87sc\") pod \"cert-manager-86cb77c54b-cpjql\" (UID: \"8ca8d146-bf33-4f3e-94e2-a794187ee78e\") " pod="cert-manager/cert-manager-86cb77c54b-cpjql" Dec 06 07:11:14 crc kubenswrapper[4945]: I1206 07:11:14.082725 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8ca8d146-bf33-4f3e-94e2-a794187ee78e-bound-sa-token\") pod \"cert-manager-86cb77c54b-cpjql\" (UID: \"8ca8d146-bf33-4f3e-94e2-a794187ee78e\") " pod="cert-manager/cert-manager-86cb77c54b-cpjql" Dec 06 07:11:14 crc kubenswrapper[4945]: I1206 07:11:14.100867 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8ca8d146-bf33-4f3e-94e2-a794187ee78e-bound-sa-token\") pod \"cert-manager-86cb77c54b-cpjql\" (UID: \"8ca8d146-bf33-4f3e-94e2-a794187ee78e\") " pod="cert-manager/cert-manager-86cb77c54b-cpjql" Dec 06 07:11:14 crc kubenswrapper[4945]: I1206 07:11:14.101351 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q87sc\" (UniqueName: \"kubernetes.io/projected/8ca8d146-bf33-4f3e-94e2-a794187ee78e-kube-api-access-q87sc\") pod \"cert-manager-86cb77c54b-cpjql\" (UID: \"8ca8d146-bf33-4f3e-94e2-a794187ee78e\") " pod="cert-manager/cert-manager-86cb77c54b-cpjql" Dec 06 07:11:14 crc kubenswrapper[4945]: I1206 07:11:14.150349 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-cpjql" Dec 06 07:11:14 crc kubenswrapper[4945]: I1206 07:11:14.583178 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-cpjql"] Dec 06 07:11:15 crc kubenswrapper[4945]: I1206 07:11:15.007400 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-ms6cl" Dec 06 07:11:15 crc kubenswrapper[4945]: I1206 07:11:15.401837 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-cpjql" event={"ID":"8ca8d146-bf33-4f3e-94e2-a794187ee78e","Type":"ContainerStarted","Data":"1f894efd68559685d5ccdbdef5fbd3dec331d9d1dd7b60ffa02e1e731b4ffd6f"} Dec 06 07:11:15 crc kubenswrapper[4945]: I1206 07:11:15.402338 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-cpjql" event={"ID":"8ca8d146-bf33-4f3e-94e2-a794187ee78e","Type":"ContainerStarted","Data":"4508abb6478de5fc4bcb91267d5edb16c02bdba08d6d2747e424a1e66fba244a"} Dec 06 07:11:15 crc kubenswrapper[4945]: I1206 07:11:15.420498 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-cpjql" podStartSLOduration=2.42047638 podStartE2EDuration="2.42047638s" podCreationTimestamp="2025-12-06 07:11:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:11:15.419602256 +0000 UTC m=+1088.874463310" watchObservedRunningTime="2025-12-06 07:11:15.42047638 +0000 UTC m=+1088.875337424" Dec 06 07:11:18 crc kubenswrapper[4945]: I1206 07:11:18.340169 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-s7n6v"] Dec 06 07:11:18 crc kubenswrapper[4945]: I1206 07:11:18.341235 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s7n6v" Dec 06 07:11:18 crc kubenswrapper[4945]: W1206 07:11:18.343068 4945 reflector.go:561] object-"openstack-operators"/"openstack-operator-index-dockercfg-pkslh": failed to list *v1.Secret: secrets "openstack-operator-index-dockercfg-pkslh" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Dec 06 07:11:18 crc kubenswrapper[4945]: E1206 07:11:18.343137 4945 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"openstack-operator-index-dockercfg-pkslh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openstack-operator-index-dockercfg-pkslh\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 07:11:18 crc kubenswrapper[4945]: I1206 07:11:18.346432 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 06 07:11:18 crc kubenswrapper[4945]: I1206 07:11:18.351489 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 06 07:11:18 crc kubenswrapper[4945]: I1206 07:11:18.399822 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-s7n6v"] Dec 06 07:11:18 crc kubenswrapper[4945]: I1206 07:11:18.443091 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tbfh\" (UniqueName: \"kubernetes.io/projected/8761a21b-a781-4f56-8eea-468028bce4a3-kube-api-access-8tbfh\") pod \"openstack-operator-index-s7n6v\" (UID: \"8761a21b-a781-4f56-8eea-468028bce4a3\") " pod="openstack-operators/openstack-operator-index-s7n6v" Dec 06 07:11:18 crc kubenswrapper[4945]: I1206 07:11:18.545085 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tbfh\" (UniqueName: \"kubernetes.io/projected/8761a21b-a781-4f56-8eea-468028bce4a3-kube-api-access-8tbfh\") pod \"openstack-operator-index-s7n6v\" (UID: \"8761a21b-a781-4f56-8eea-468028bce4a3\") " pod="openstack-operators/openstack-operator-index-s7n6v" Dec 06 07:11:18 crc kubenswrapper[4945]: I1206 07:11:18.565369 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tbfh\" (UniqueName: \"kubernetes.io/projected/8761a21b-a781-4f56-8eea-468028bce4a3-kube-api-access-8tbfh\") pod \"openstack-operator-index-s7n6v\" (UID: \"8761a21b-a781-4f56-8eea-468028bce4a3\") " pod="openstack-operators/openstack-operator-index-s7n6v" Dec 06 07:11:19 crc kubenswrapper[4945]: I1206 07:11:19.500863 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-pkslh" Dec 06 07:11:19 crc kubenswrapper[4945]: I1206 07:11:19.502519 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s7n6v" Dec 06 07:11:19 crc kubenswrapper[4945]: I1206 07:11:19.938413 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-s7n6v"] Dec 06 07:11:20 crc kubenswrapper[4945]: I1206 07:11:20.440662 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s7n6v" event={"ID":"8761a21b-a781-4f56-8eea-468028bce4a3","Type":"ContainerStarted","Data":"a1d15cab4e537c27f0039afc208047eda89e97095b17ebef8143e5f291813a0e"} Dec 06 07:11:21 crc kubenswrapper[4945]: I1206 07:11:21.914173 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-s7n6v"] Dec 06 07:11:22 crc kubenswrapper[4945]: I1206 07:11:22.456682 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s7n6v" event={"ID":"8761a21b-a781-4f56-8eea-468028bce4a3","Type":"ContainerStarted","Data":"081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b"} Dec 06 07:11:22 crc kubenswrapper[4945]: I1206 07:11:22.456899 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-s7n6v" podUID="8761a21b-a781-4f56-8eea-468028bce4a3" containerName="registry-server" containerID="cri-o://081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b" gracePeriod=2 Dec 06 07:11:22 crc kubenswrapper[4945]: I1206 07:11:22.478452 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-s7n6v" podStartSLOduration=2.571326332 podStartE2EDuration="4.478422043s" podCreationTimestamp="2025-12-06 07:11:18 +0000 UTC" firstStartedPulling="2025-12-06 07:11:19.94597351 +0000 UTC m=+1093.400834554" lastFinishedPulling="2025-12-06 07:11:21.853069221 +0000 UTC m=+1095.307930265" observedRunningTime="2025-12-06 07:11:22.471752043 +0000 UTC m=+1095.926613097" watchObservedRunningTime="2025-12-06 07:11:22.478422043 +0000 UTC m=+1095.933283107" Dec 06 07:11:22 crc kubenswrapper[4945]: I1206 07:11:22.716239 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-5h9ss"] Dec 06 07:11:22 crc kubenswrapper[4945]: I1206 07:11:22.717096 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5h9ss" Dec 06 07:11:22 crc kubenswrapper[4945]: I1206 07:11:22.728250 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5h9ss"] Dec 06 07:11:22 crc kubenswrapper[4945]: I1206 07:11:22.811127 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qlnx\" (UniqueName: \"kubernetes.io/projected/5da80984-5ba5-48be-b853-847ae00cba6c-kube-api-access-5qlnx\") pod \"openstack-operator-index-5h9ss\" (UID: \"5da80984-5ba5-48be-b853-847ae00cba6c\") " pod="openstack-operators/openstack-operator-index-5h9ss" Dec 06 07:11:22 crc kubenswrapper[4945]: I1206 07:11:22.911951 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qlnx\" (UniqueName: \"kubernetes.io/projected/5da80984-5ba5-48be-b853-847ae00cba6c-kube-api-access-5qlnx\") pod \"openstack-operator-index-5h9ss\" (UID: \"5da80984-5ba5-48be-b853-847ae00cba6c\") " pod="openstack-operators/openstack-operator-index-5h9ss" Dec 06 07:11:22 crc kubenswrapper[4945]: I1206 07:11:22.934462 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qlnx\" (UniqueName: \"kubernetes.io/projected/5da80984-5ba5-48be-b853-847ae00cba6c-kube-api-access-5qlnx\") pod \"openstack-operator-index-5h9ss\" (UID: \"5da80984-5ba5-48be-b853-847ae00cba6c\") " pod="openstack-operators/openstack-operator-index-5h9ss" Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.033030 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5h9ss" Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.033440 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s7n6v" Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.115001 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tbfh\" (UniqueName: \"kubernetes.io/projected/8761a21b-a781-4f56-8eea-468028bce4a3-kube-api-access-8tbfh\") pod \"8761a21b-a781-4f56-8eea-468028bce4a3\" (UID: \"8761a21b-a781-4f56-8eea-468028bce4a3\") " Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.119789 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8761a21b-a781-4f56-8eea-468028bce4a3-kube-api-access-8tbfh" (OuterVolumeSpecName: "kube-api-access-8tbfh") pod "8761a21b-a781-4f56-8eea-468028bce4a3" (UID: "8761a21b-a781-4f56-8eea-468028bce4a3"). InnerVolumeSpecName "kube-api-access-8tbfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.219582 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tbfh\" (UniqueName: \"kubernetes.io/projected/8761a21b-a781-4f56-8eea-468028bce4a3-kube-api-access-8tbfh\") on node \"crc\" DevicePath \"\"" Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.465995 4945 generic.go:334] "Generic (PLEG): container finished" podID="8761a21b-a781-4f56-8eea-468028bce4a3" containerID="081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b" exitCode=0 Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.466062 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s7n6v" Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.466047 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s7n6v" event={"ID":"8761a21b-a781-4f56-8eea-468028bce4a3","Type":"ContainerDied","Data":"081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b"} Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.466559 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5h9ss"] Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.466590 4945 scope.go:117] "RemoveContainer" containerID="081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b" Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.466773 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s7n6v" event={"ID":"8761a21b-a781-4f56-8eea-468028bce4a3","Type":"ContainerDied","Data":"a1d15cab4e537c27f0039afc208047eda89e97095b17ebef8143e5f291813a0e"} Dec 06 07:11:23 crc kubenswrapper[4945]: W1206 07:11:23.468453 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5da80984_5ba5_48be_b853_847ae00cba6c.slice/crio-4701b3861fa773fb1d5f35cdb216484795288023dc3cec2faea73d8591466a8d WatchSource:0}: Error finding container 4701b3861fa773fb1d5f35cdb216484795288023dc3cec2faea73d8591466a8d: Status 404 returned error can't find the container with id 4701b3861fa773fb1d5f35cdb216484795288023dc3cec2faea73d8591466a8d Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.485036 4945 scope.go:117] "RemoveContainer" containerID="081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b" Dec 06 07:11:23 crc kubenswrapper[4945]: E1206 07:11:23.485546 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b\": container with ID starting with 081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b not found: ID does not exist" containerID="081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b" Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.485599 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b"} err="failed to get container status \"081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b\": rpc error: code = NotFound desc = could not find container \"081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b\": container with ID starting with 081c3e5b0dc191bb07dd904e1ae195afe3343176be8f96de471c041536e20c8b not found: ID does not exist" Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.495198 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-s7n6v"] Dec 06 07:11:23 crc kubenswrapper[4945]: I1206 07:11:23.499521 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-s7n6v"] Dec 06 07:11:24 crc kubenswrapper[4945]: I1206 07:11:24.474790 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5h9ss" event={"ID":"5da80984-5ba5-48be-b853-847ae00cba6c","Type":"ContainerStarted","Data":"4701b3861fa773fb1d5f35cdb216484795288023dc3cec2faea73d8591466a8d"} Dec 06 07:11:24 crc kubenswrapper[4945]: I1206 07:11:24.960041 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8761a21b-a781-4f56-8eea-468028bce4a3" path="/var/lib/kubelet/pods/8761a21b-a781-4f56-8eea-468028bce4a3/volumes" Dec 06 07:11:26 crc kubenswrapper[4945]: I1206 07:11:26.494861 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5h9ss" event={"ID":"5da80984-5ba5-48be-b853-847ae00cba6c","Type":"ContainerStarted","Data":"c61ea9b6678fe58ce131e0c147411aada88d98a19ddbbb92a3a2a3ba60a8a527"} Dec 06 07:11:26 crc kubenswrapper[4945]: I1206 07:11:26.508819 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-5h9ss" podStartSLOduration=2.5777769299999997 podStartE2EDuration="4.508802514s" podCreationTimestamp="2025-12-06 07:11:22 +0000 UTC" firstStartedPulling="2025-12-06 07:11:23.472066035 +0000 UTC m=+1096.926927079" lastFinishedPulling="2025-12-06 07:11:25.403091619 +0000 UTC m=+1098.857952663" observedRunningTime="2025-12-06 07:11:26.508570937 +0000 UTC m=+1099.963431991" watchObservedRunningTime="2025-12-06 07:11:26.508802514 +0000 UTC m=+1099.963663558" Dec 06 07:11:33 crc kubenswrapper[4945]: I1206 07:11:33.034004 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-5h9ss" Dec 06 07:11:33 crc kubenswrapper[4945]: I1206 07:11:33.034502 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-5h9ss" Dec 06 07:11:33 crc kubenswrapper[4945]: I1206 07:11:33.061788 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-5h9ss" Dec 06 07:11:33 crc kubenswrapper[4945]: I1206 07:11:33.565910 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-5h9ss" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.352766 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx"] Dec 06 07:11:38 crc kubenswrapper[4945]: E1206 07:11:38.353409 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8761a21b-a781-4f56-8eea-468028bce4a3" containerName="registry-server" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.353426 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8761a21b-a781-4f56-8eea-468028bce4a3" containerName="registry-server" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.353583 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8761a21b-a781-4f56-8eea-468028bce4a3" containerName="registry-server" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.354639 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.356468 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-phpsm" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.364104 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx"] Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.513620 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.513825 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dknq5\" (UniqueName: \"kubernetes.io/projected/9329d400-df2a-48e9-bdee-cdaa2304ae0e-kube-api-access-dknq5\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.513925 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.616032 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.616149 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dknq5\" (UniqueName: \"kubernetes.io/projected/9329d400-df2a-48e9-bdee-cdaa2304ae0e-kube-api-access-dknq5\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.616214 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.616934 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.617144 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.637250 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dknq5\" (UniqueName: \"kubernetes.io/projected/9329d400-df2a-48e9-bdee-cdaa2304ae0e-kube-api-access-dknq5\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:38 crc kubenswrapper[4945]: I1206 07:11:38.671328 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:39 crc kubenswrapper[4945]: I1206 07:11:39.070463 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx"] Dec 06 07:11:39 crc kubenswrapper[4945]: W1206 07:11:39.074303 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9329d400_df2a_48e9_bdee_cdaa2304ae0e.slice/crio-3d036e8cc99d1a4cf74b096d6b67b20c6b1f705d42b83cb876ce091f77997cc8 WatchSource:0}: Error finding container 3d036e8cc99d1a4cf74b096d6b67b20c6b1f705d42b83cb876ce091f77997cc8: Status 404 returned error can't find the container with id 3d036e8cc99d1a4cf74b096d6b67b20c6b1f705d42b83cb876ce091f77997cc8 Dec 06 07:11:39 crc kubenswrapper[4945]: I1206 07:11:39.585687 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" event={"ID":"9329d400-df2a-48e9-bdee-cdaa2304ae0e","Type":"ContainerStarted","Data":"3d036e8cc99d1a4cf74b096d6b67b20c6b1f705d42b83cb876ce091f77997cc8"} Dec 06 07:11:40 crc kubenswrapper[4945]: I1206 07:11:40.593962 4945 generic.go:334] "Generic (PLEG): container finished" podID="9329d400-df2a-48e9-bdee-cdaa2304ae0e" containerID="879508bc98ba03ffd14035676b3ccd3316e9310f45fdcd14da4d50370cc77e96" exitCode=0 Dec 06 07:11:40 crc kubenswrapper[4945]: I1206 07:11:40.594011 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" event={"ID":"9329d400-df2a-48e9-bdee-cdaa2304ae0e","Type":"ContainerDied","Data":"879508bc98ba03ffd14035676b3ccd3316e9310f45fdcd14da4d50370cc77e96"} Dec 06 07:11:44 crc kubenswrapper[4945]: I1206 07:11:44.625338 4945 generic.go:334] "Generic (PLEG): container finished" podID="9329d400-df2a-48e9-bdee-cdaa2304ae0e" containerID="b7674a3ddc75e00171116baea1edc19bf9d20765afb1d1a22a86e9e8f6bb6d8c" exitCode=0 Dec 06 07:11:44 crc kubenswrapper[4945]: I1206 07:11:44.625887 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" event={"ID":"9329d400-df2a-48e9-bdee-cdaa2304ae0e","Type":"ContainerDied","Data":"b7674a3ddc75e00171116baea1edc19bf9d20765afb1d1a22a86e9e8f6bb6d8c"} Dec 06 07:11:45 crc kubenswrapper[4945]: I1206 07:11:45.634459 4945 generic.go:334] "Generic (PLEG): container finished" podID="9329d400-df2a-48e9-bdee-cdaa2304ae0e" containerID="dce94c81d0c2f463c5ec079ed0ec79e7af92cc5f462e25cffe315faf8967f689" exitCode=0 Dec 06 07:11:45 crc kubenswrapper[4945]: I1206 07:11:45.634564 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" event={"ID":"9329d400-df2a-48e9-bdee-cdaa2304ae0e","Type":"ContainerDied","Data":"dce94c81d0c2f463c5ec079ed0ec79e7af92cc5f462e25cffe315faf8967f689"} Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.110792 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.230212 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-bundle\") pod \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.230310 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-util\") pod \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.230464 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dknq5\" (UniqueName: \"kubernetes.io/projected/9329d400-df2a-48e9-bdee-cdaa2304ae0e-kube-api-access-dknq5\") pod \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\" (UID: \"9329d400-df2a-48e9-bdee-cdaa2304ae0e\") " Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.231198 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-bundle" (OuterVolumeSpecName: "bundle") pod "9329d400-df2a-48e9-bdee-cdaa2304ae0e" (UID: "9329d400-df2a-48e9-bdee-cdaa2304ae0e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.235246 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9329d400-df2a-48e9-bdee-cdaa2304ae0e-kube-api-access-dknq5" (OuterVolumeSpecName: "kube-api-access-dknq5") pod "9329d400-df2a-48e9-bdee-cdaa2304ae0e" (UID: "9329d400-df2a-48e9-bdee-cdaa2304ae0e"). InnerVolumeSpecName "kube-api-access-dknq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.242074 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-util" (OuterVolumeSpecName: "util") pod "9329d400-df2a-48e9-bdee-cdaa2304ae0e" (UID: "9329d400-df2a-48e9-bdee-cdaa2304ae0e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.332610 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.332703 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9329d400-df2a-48e9-bdee-cdaa2304ae0e-util\") on node \"crc\" DevicePath \"\"" Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.332716 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dknq5\" (UniqueName: \"kubernetes.io/projected/9329d400-df2a-48e9-bdee-cdaa2304ae0e-kube-api-access-dknq5\") on node \"crc\" DevicePath \"\"" Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.647616 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" event={"ID":"9329d400-df2a-48e9-bdee-cdaa2304ae0e","Type":"ContainerDied","Data":"3d036e8cc99d1a4cf74b096d6b67b20c6b1f705d42b83cb876ce091f77997cc8"} Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.647652 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d036e8cc99d1a4cf74b096d6b67b20c6b1f705d42b83cb876ce091f77997cc8" Dec 06 07:11:47 crc kubenswrapper[4945]: I1206 07:11:47.647718 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.657036 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5"] Dec 06 07:11:50 crc kubenswrapper[4945]: E1206 07:11:50.657580 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9329d400-df2a-48e9-bdee-cdaa2304ae0e" containerName="pull" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.657594 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9329d400-df2a-48e9-bdee-cdaa2304ae0e" containerName="pull" Dec 06 07:11:50 crc kubenswrapper[4945]: E1206 07:11:50.657615 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9329d400-df2a-48e9-bdee-cdaa2304ae0e" containerName="extract" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.657621 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9329d400-df2a-48e9-bdee-cdaa2304ae0e" containerName="extract" Dec 06 07:11:50 crc kubenswrapper[4945]: E1206 07:11:50.657633 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9329d400-df2a-48e9-bdee-cdaa2304ae0e" containerName="util" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.657639 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9329d400-df2a-48e9-bdee-cdaa2304ae0e" containerName="util" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.657759 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9329d400-df2a-48e9-bdee-cdaa2304ae0e" containerName="extract" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.658167 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.660955 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-rftgs" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.711028 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5"] Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.737550 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ggg6\" (UniqueName: \"kubernetes.io/projected/5360a460-56fc-4dd6-9014-614ed5a8e05e-kube-api-access-2ggg6\") pod \"openstack-operator-controller-operator-55b6fb9447-lfnj5\" (UID: \"5360a460-56fc-4dd6-9014-614ed5a8e05e\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.838932 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ggg6\" (UniqueName: \"kubernetes.io/projected/5360a460-56fc-4dd6-9014-614ed5a8e05e-kube-api-access-2ggg6\") pod \"openstack-operator-controller-operator-55b6fb9447-lfnj5\" (UID: \"5360a460-56fc-4dd6-9014-614ed5a8e05e\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.892403 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ggg6\" (UniqueName: \"kubernetes.io/projected/5360a460-56fc-4dd6-9014-614ed5a8e05e-kube-api-access-2ggg6\") pod \"openstack-operator-controller-operator-55b6fb9447-lfnj5\" (UID: \"5360a460-56fc-4dd6-9014-614ed5a8e05e\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5" Dec 06 07:11:50 crc kubenswrapper[4945]: I1206 07:11:50.974148 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5" Dec 06 07:11:51 crc kubenswrapper[4945]: I1206 07:11:51.255889 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5"] Dec 06 07:11:51 crc kubenswrapper[4945]: I1206 07:11:51.775846 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5" event={"ID":"5360a460-56fc-4dd6-9014-614ed5a8e05e","Type":"ContainerStarted","Data":"7ac395bb8accf17e8e78da58475daa53d81c737fb3b19b6a1e13eeb7398a6bb4"} Dec 06 07:11:58 crc kubenswrapper[4945]: I1206 07:11:58.918947 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5" event={"ID":"5360a460-56fc-4dd6-9014-614ed5a8e05e","Type":"ContainerStarted","Data":"37a082c9c607cf7db9b9617477fa3d0fea0615d7dd30e749b2ad0a425f420cb0"} Dec 06 07:11:58 crc kubenswrapper[4945]: I1206 07:11:58.920121 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5" Dec 06 07:11:58 crc kubenswrapper[4945]: I1206 07:11:58.950625 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5" podStartSLOduration=2.233185042 podStartE2EDuration="8.950604497s" podCreationTimestamp="2025-12-06 07:11:50 +0000 UTC" firstStartedPulling="2025-12-06 07:11:51.259532925 +0000 UTC m=+1124.714393969" lastFinishedPulling="2025-12-06 07:11:57.97695238 +0000 UTC m=+1131.431813424" observedRunningTime="2025-12-06 07:11:58.946939959 +0000 UTC m=+1132.401801033" watchObservedRunningTime="2025-12-06 07:11:58.950604497 +0000 UTC m=+1132.405465541" Dec 06 07:12:10 crc kubenswrapper[4945]: I1206 07:12:10.977627 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lfnj5" Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.879018 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g"] Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.880883 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.884453 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-dlnhn" Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.885468 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f"] Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.886746 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.888227 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-hgvh9" Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.893997 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g"] Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.924180 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f"] Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.929441 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4"] Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.930495 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.937702 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-wrxsf" Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.950382 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4"] Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.958709 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh"] Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.959830 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.961639 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-jn7jg" Dec 06 07:12:31 crc kubenswrapper[4945]: I1206 07:12:31.976837 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.022879 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.024085 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.029947 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-ltfbw" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.045718 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq7fp\" (UniqueName: \"kubernetes.io/projected/afd67b97-8d07-4522-b37d-019abf73ee85-kube-api-access-mq7fp\") pod \"cinder-operator-controller-manager-859b6ccc6-qd55f\" (UID: \"afd67b97-8d07-4522-b37d-019abf73ee85\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.045835 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7srq\" (UniqueName: \"kubernetes.io/projected/7305df73-ea7d-43f5-9119-9c12942796bc-kube-api-access-s7srq\") pod \"barbican-operator-controller-manager-7d9dfd778-wqq5g\" (UID: \"7305df73-ea7d-43f5-9119-9c12942796bc\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.045889 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhb4h\" (UniqueName: \"kubernetes.io/projected/31beec86-b14f-43c6-9307-a776454ed9b7-kube-api-access-qhb4h\") pod \"designate-operator-controller-manager-78b4bc895b-kjxk4\" (UID: \"31beec86-b14f-43c6-9307-a776454ed9b7\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.049352 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.050340 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.053696 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-9shxf" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.062594 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.065885 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.067130 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.075702 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-2sp6d" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.075903 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.076024 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.089072 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.111571 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.112698 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.116086 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-kkv9m" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.120853 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.122053 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.142361 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.143731 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.146277 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-s767c" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.147177 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-w2srb" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.149131 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smt2f\" (UniqueName: \"kubernetes.io/projected/c925c68e-4099-403e-b8ba-84d2d9bc689e-kube-api-access-smt2f\") pod \"glance-operator-controller-manager-77987cd8cd-cgbvh\" (UID: \"c925c68e-4099-403e-b8ba-84d2d9bc689e\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.149196 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4g2r\" (UniqueName: \"kubernetes.io/projected/f32f27d3-90de-4c9e-8664-68e149e42ffc-kube-api-access-l4g2r\") pod \"horizon-operator-controller-manager-68c6d99b8f-cchx8\" (UID: \"f32f27d3-90de-4c9e-8664-68e149e42ffc\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.149247 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhb4h\" (UniqueName: \"kubernetes.io/projected/31beec86-b14f-43c6-9307-a776454ed9b7-kube-api-access-qhb4h\") pod \"designate-operator-controller-manager-78b4bc895b-kjxk4\" (UID: \"31beec86-b14f-43c6-9307-a776454ed9b7\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.149347 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq7fp\" (UniqueName: \"kubernetes.io/projected/afd67b97-8d07-4522-b37d-019abf73ee85-kube-api-access-mq7fp\") pod \"cinder-operator-controller-manager-859b6ccc6-qd55f\" (UID: \"afd67b97-8d07-4522-b37d-019abf73ee85\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.149402 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7srq\" (UniqueName: \"kubernetes.io/projected/7305df73-ea7d-43f5-9119-9c12942796bc-kube-api-access-s7srq\") pod \"barbican-operator-controller-manager-7d9dfd778-wqq5g\" (UID: \"7305df73-ea7d-43f5-9119-9c12942796bc\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.164698 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.172512 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.186425 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq7fp\" (UniqueName: \"kubernetes.io/projected/afd67b97-8d07-4522-b37d-019abf73ee85-kube-api-access-mq7fp\") pod \"cinder-operator-controller-manager-859b6ccc6-qd55f\" (UID: \"afd67b97-8d07-4522-b37d-019abf73ee85\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.186551 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.200003 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhb4h\" (UniqueName: \"kubernetes.io/projected/31beec86-b14f-43c6-9307-a776454ed9b7-kube-api-access-qhb4h\") pod \"designate-operator-controller-manager-78b4bc895b-kjxk4\" (UID: \"31beec86-b14f-43c6-9307-a776454ed9b7\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.201448 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.203032 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.205807 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-w2ldv" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.227528 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.228203 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.237494 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.238818 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.244491 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.245932 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.247665 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-szm4j" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.249072 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7srq\" (UniqueName: \"kubernetes.io/projected/7305df73-ea7d-43f5-9119-9c12942796bc-kube-api-access-s7srq\") pod \"barbican-operator-controller-manager-7d9dfd778-wqq5g\" (UID: \"7305df73-ea7d-43f5-9119-9c12942796bc\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252459 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smt2f\" (UniqueName: \"kubernetes.io/projected/c925c68e-4099-403e-b8ba-84d2d9bc689e-kube-api-access-smt2f\") pod \"glance-operator-controller-manager-77987cd8cd-cgbvh\" (UID: \"c925c68e-4099-403e-b8ba-84d2d9bc689e\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252516 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4g2r\" (UniqueName: \"kubernetes.io/projected/f32f27d3-90de-4c9e-8664-68e149e42ffc-kube-api-access-l4g2r\") pod \"horizon-operator-controller-manager-68c6d99b8f-cchx8\" (UID: \"f32f27d3-90de-4c9e-8664-68e149e42ffc\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252563 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmbw6\" (UniqueName: \"kubernetes.io/projected/40318b4d-ed4f-478b-8333-8e1cb0729b34-kube-api-access-nmbw6\") pod \"manila-operator-controller-manager-7c79b5df47-p2gzt\" (UID: \"40318b4d-ed4f-478b-8333-8e1cb0729b34\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252602 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmgz7\" (UniqueName: \"kubernetes.io/projected/8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd-kube-api-access-zmgz7\") pod \"heat-operator-controller-manager-5f64f6f8bb-n8gdh\" (UID: \"8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252620 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2xcd\" (UniqueName: \"kubernetes.io/projected/db9c0007-43e1-4cb2-ae8a-44a4731254b7-kube-api-access-h2xcd\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252648 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkmrf\" (UniqueName: \"kubernetes.io/projected/6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc-kube-api-access-tkmrf\") pod \"nova-operator-controller-manager-697bc559fc-bpsxr\" (UID: \"6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252690 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x54sz\" (UniqueName: \"kubernetes.io/projected/7c3f47b7-f1f2-408d-80d0-03a0ea9286ea-kube-api-access-x54sz\") pod \"keystone-operator-controller-manager-7765d96ddf-x98dd\" (UID: \"7c3f47b7-f1f2-408d-80d0-03a0ea9286ea\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252743 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgsw2\" (UniqueName: \"kubernetes.io/projected/0e92db58-a1fc-4035-b635-f274d7411813-kube-api-access-bgsw2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-5njm2\" (UID: \"0e92db58-a1fc-4035-b635-f274d7411813\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252774 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n95tc\" (UniqueName: \"kubernetes.io/projected/3c4830bb-bc9c-447e-bb30-9c481fae8234-kube-api-access-n95tc\") pod \"mariadb-operator-controller-manager-56bbcc9d85-f6rz7\" (UID: \"3c4830bb-bc9c-447e-bb30-9c481fae8234\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252804 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf2jv\" (UniqueName: \"kubernetes.io/projected/b89bffc9-f64a-4d0b-b5e5-104db0d681dd-kube-api-access-wf2jv\") pod \"ironic-operator-controller-manager-6c548fd776-r7zm6\" (UID: \"b89bffc9-f64a-4d0b-b5e5-104db0d681dd\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.252836 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.255192 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-lpb9g" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.354247 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.383616 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf2jv\" (UniqueName: \"kubernetes.io/projected/b89bffc9-f64a-4d0b-b5e5-104db0d681dd-kube-api-access-wf2jv\") pod \"ironic-operator-controller-manager-6c548fd776-r7zm6\" (UID: \"b89bffc9-f64a-4d0b-b5e5-104db0d681dd\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.383697 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.383906 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmbw6\" (UniqueName: \"kubernetes.io/projected/40318b4d-ed4f-478b-8333-8e1cb0729b34-kube-api-access-nmbw6\") pod \"manila-operator-controller-manager-7c79b5df47-p2gzt\" (UID: \"40318b4d-ed4f-478b-8333-8e1cb0729b34\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.383988 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmgz7\" (UniqueName: \"kubernetes.io/projected/8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd-kube-api-access-zmgz7\") pod \"heat-operator-controller-manager-5f64f6f8bb-n8gdh\" (UID: \"8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.384022 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2xcd\" (UniqueName: \"kubernetes.io/projected/db9c0007-43e1-4cb2-ae8a-44a4731254b7-kube-api-access-h2xcd\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.384117 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkmrf\" (UniqueName: \"kubernetes.io/projected/6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc-kube-api-access-tkmrf\") pod \"nova-operator-controller-manager-697bc559fc-bpsxr\" (UID: \"6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.384351 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x54sz\" (UniqueName: \"kubernetes.io/projected/7c3f47b7-f1f2-408d-80d0-03a0ea9286ea-kube-api-access-x54sz\") pod \"keystone-operator-controller-manager-7765d96ddf-x98dd\" (UID: \"7c3f47b7-f1f2-408d-80d0-03a0ea9286ea\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.388027 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgsw2\" (UniqueName: \"kubernetes.io/projected/0e92db58-a1fc-4035-b635-f274d7411813-kube-api-access-bgsw2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-5njm2\" (UID: \"0e92db58-a1fc-4035-b635-f274d7411813\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.388076 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n95tc\" (UniqueName: \"kubernetes.io/projected/3c4830bb-bc9c-447e-bb30-9c481fae8234-kube-api-access-n95tc\") pod \"mariadb-operator-controller-manager-56bbcc9d85-f6rz7\" (UID: \"3c4830bb-bc9c-447e-bb30-9c481fae8234\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" Dec 06 07:12:32 crc kubenswrapper[4945]: E1206 07:12:32.401101 4945 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 07:12:32 crc kubenswrapper[4945]: E1206 07:12:32.401178 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert podName:db9c0007-43e1-4cb2-ae8a-44a4731254b7 nodeName:}" failed. No retries permitted until 2025-12-06 07:12:32.901154419 +0000 UTC m=+1166.356015463 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert") pod "infra-operator-controller-manager-57548d458d-qhvjz" (UID: "db9c0007-43e1-4cb2-ae8a-44a4731254b7") : secret "infra-operator-webhook-server-cert" not found Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.429162 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4g2r\" (UniqueName: \"kubernetes.io/projected/f32f27d3-90de-4c9e-8664-68e149e42ffc-kube-api-access-l4g2r\") pod \"horizon-operator-controller-manager-68c6d99b8f-cchx8\" (UID: \"f32f27d3-90de-4c9e-8664-68e149e42ffc\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.476138 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgsw2\" (UniqueName: \"kubernetes.io/projected/0e92db58-a1fc-4035-b635-f274d7411813-kube-api-access-bgsw2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-5njm2\" (UID: \"0e92db58-a1fc-4035-b635-f274d7411813\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.477753 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmgz7\" (UniqueName: \"kubernetes.io/projected/8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd-kube-api-access-zmgz7\") pod \"heat-operator-controller-manager-5f64f6f8bb-n8gdh\" (UID: \"8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.480826 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmbw6\" (UniqueName: \"kubernetes.io/projected/40318b4d-ed4f-478b-8333-8e1cb0729b34-kube-api-access-nmbw6\") pod \"manila-operator-controller-manager-7c79b5df47-p2gzt\" (UID: \"40318b4d-ed4f-478b-8333-8e1cb0729b34\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.482963 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n95tc\" (UniqueName: \"kubernetes.io/projected/3c4830bb-bc9c-447e-bb30-9c481fae8234-kube-api-access-n95tc\") pod \"mariadb-operator-controller-manager-56bbcc9d85-f6rz7\" (UID: \"3c4830bb-bc9c-447e-bb30-9c481fae8234\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.483566 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x54sz\" (UniqueName: \"kubernetes.io/projected/7c3f47b7-f1f2-408d-80d0-03a0ea9286ea-kube-api-access-x54sz\") pod \"keystone-operator-controller-manager-7765d96ddf-x98dd\" (UID: \"7c3f47b7-f1f2-408d-80d0-03a0ea9286ea\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.486345 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.486540 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkmrf\" (UniqueName: \"kubernetes.io/projected/6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc-kube-api-access-tkmrf\") pod \"nova-operator-controller-manager-697bc559fc-bpsxr\" (UID: \"6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.486789 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2xcd\" (UniqueName: \"kubernetes.io/projected/db9c0007-43e1-4cb2-ae8a-44a4731254b7-kube-api-access-h2xcd\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.489746 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smt2f\" (UniqueName: \"kubernetes.io/projected/c925c68e-4099-403e-b8ba-84d2d9bc689e-kube-api-access-smt2f\") pod \"glance-operator-controller-manager-77987cd8cd-cgbvh\" (UID: \"c925c68e-4099-403e-b8ba-84d2d9bc689e\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.496165 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf2jv\" (UniqueName: \"kubernetes.io/projected/b89bffc9-f64a-4d0b-b5e5-104db0d681dd-kube-api-access-wf2jv\") pod \"ironic-operator-controller-manager-6c548fd776-r7zm6\" (UID: \"b89bffc9-f64a-4d0b-b5e5-104db0d681dd\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.513458 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.515451 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.534014 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.550476 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.581992 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.595838 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-pd42f"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.597817 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.601544 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-t27lb" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.627121 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-pd42f"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.650224 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.652173 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.654508 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.667902 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-zlmf9" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.668069 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.676337 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.677631 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.677730 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.687579 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-c47wl" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.739777 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.739912 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfpxt\" (UniqueName: \"kubernetes.io/projected/9bfd7a39-d83d-42de-bb30-0d1dc17bedca-kube-api-access-hfpxt\") pod \"octavia-operator-controller-manager-998648c74-pd42f\" (UID: \"9bfd7a39-d83d-42de-bb30-0d1dc17bedca\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.739937 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnx45\" (UniqueName: \"kubernetes.io/projected/6c628f21-0f62-4761-9dd1-f6377a5633a5-kube-api-access-jnx45\") pod \"ovn-operator-controller-manager-b6456fdb6-lwkxm\" (UID: \"6c628f21-0f62-4761-9dd1-f6377a5633a5\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.739959 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxj5k\" (UniqueName: \"kubernetes.io/projected/19942e24-acd1-469c-94b2-37063e88c190-kube-api-access-wxj5k\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.750662 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.756623 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.782670 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.796116 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.796332 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.804752 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.805958 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.811588 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-nbqg6" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.811770 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.824484 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.826701 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.830364 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.836671 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.841304 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.843388 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-hdsmz" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.843917 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49wsl\" (UniqueName: \"kubernetes.io/projected/4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4-kube-api-access-49wsl\") pod \"telemetry-operator-controller-manager-76cc84c6bb-c4q9q\" (UID: \"4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.843955 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfpxt\" (UniqueName: \"kubernetes.io/projected/9bfd7a39-d83d-42de-bb30-0d1dc17bedca-kube-api-access-hfpxt\") pod \"octavia-operator-controller-manager-998648c74-pd42f\" (UID: \"9bfd7a39-d83d-42de-bb30-0d1dc17bedca\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.843972 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnx45\" (UniqueName: \"kubernetes.io/projected/6c628f21-0f62-4761-9dd1-f6377a5633a5-kube-api-access-jnx45\") pod \"ovn-operator-controller-manager-b6456fdb6-lwkxm\" (UID: \"6c628f21-0f62-4761-9dd1-f6377a5633a5\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.843992 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxj5k\" (UniqueName: \"kubernetes.io/projected/19942e24-acd1-469c-94b2-37063e88c190-kube-api-access-wxj5k\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.844015 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.844075 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz8tq\" (UniqueName: \"kubernetes.io/projected/716df590-6583-4b4b-bfdc-4d6bb2ac05a4-kube-api-access-sz8tq\") pod \"swift-operator-controller-manager-5f8c65bbfc-6tvc7\" (UID: \"716df590-6583-4b4b-bfdc-4d6bb2ac05a4\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.844110 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr52l\" (UniqueName: \"kubernetes.io/projected/039d67fe-f691-49a7-8e38-4a8dc2a2227e-kube-api-access-qr52l\") pod \"placement-operator-controller-manager-78f8948974-7cn4x\" (UID: \"039d67fe-f691-49a7-8e38-4a8dc2a2227e\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" Dec 06 07:12:32 crc kubenswrapper[4945]: E1206 07:12:32.845639 4945 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 07:12:32 crc kubenswrapper[4945]: E1206 07:12:32.845678 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert podName:19942e24-acd1-469c-94b2-37063e88c190 nodeName:}" failed. No retries permitted until 2025-12-06 07:12:33.345663836 +0000 UTC m=+1166.800524870 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f56d49f" (UID: "19942e24-acd1-469c-94b2-37063e88c190") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.849350 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.851126 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-9w98t" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.864036 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.883506 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-v5krd"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.885095 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.891566 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-l4f9t" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.900273 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxj5k\" (UniqueName: \"kubernetes.io/projected/19942e24-acd1-469c-94b2-37063e88c190-kube-api-access-wxj5k\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.906340 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.907481 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.910356 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-v5krd"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.931120 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfpxt\" (UniqueName: \"kubernetes.io/projected/9bfd7a39-d83d-42de-bb30-0d1dc17bedca-kube-api-access-hfpxt\") pod \"octavia-operator-controller-manager-998648c74-pd42f\" (UID: \"9bfd7a39-d83d-42de-bb30-0d1dc17bedca\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.931188 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx"] Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.933480 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-4glbp" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.934546 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnx45\" (UniqueName: \"kubernetes.io/projected/6c628f21-0f62-4761-9dd1-f6377a5633a5-kube-api-access-jnx45\") pod \"ovn-operator-controller-manager-b6456fdb6-lwkxm\" (UID: \"6c628f21-0f62-4761-9dd1-f6377a5633a5\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.940902 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.958426 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz8tq\" (UniqueName: \"kubernetes.io/projected/716df590-6583-4b4b-bfdc-4d6bb2ac05a4-kube-api-access-sz8tq\") pod \"swift-operator-controller-manager-5f8c65bbfc-6tvc7\" (UID: \"716df590-6583-4b4b-bfdc-4d6bb2ac05a4\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.958497 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr52l\" (UniqueName: \"kubernetes.io/projected/039d67fe-f691-49a7-8e38-4a8dc2a2227e-kube-api-access-qr52l\") pod \"placement-operator-controller-manager-78f8948974-7cn4x\" (UID: \"039d67fe-f691-49a7-8e38-4a8dc2a2227e\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.958536 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.958565 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49wsl\" (UniqueName: \"kubernetes.io/projected/4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4-kube-api-access-49wsl\") pod \"telemetry-operator-controller-manager-76cc84c6bb-c4q9q\" (UID: \"4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.958636 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blfjj\" (UniqueName: \"kubernetes.io/projected/f19785e5-2908-48b0-aa6e-ea4f37bbd506-kube-api-access-blfjj\") pod \"test-operator-controller-manager-5854674fcc-v5krd\" (UID: \"f19785e5-2908-48b0-aa6e-ea4f37bbd506\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" Dec 06 07:12:32 crc kubenswrapper[4945]: E1206 07:12:32.960816 4945 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 07:12:32 crc kubenswrapper[4945]: E1206 07:12:32.960900 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert podName:db9c0007-43e1-4cb2-ae8a-44a4731254b7 nodeName:}" failed. No retries permitted until 2025-12-06 07:12:33.960883613 +0000 UTC m=+1167.415744657 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert") pod "infra-operator-controller-manager-57548d458d-qhvjz" (UID: "db9c0007-43e1-4cb2-ae8a-44a4731254b7") : secret "infra-operator-webhook-server-cert" not found Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.989779 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr52l\" (UniqueName: \"kubernetes.io/projected/039d67fe-f691-49a7-8e38-4a8dc2a2227e-kube-api-access-qr52l\") pod \"placement-operator-controller-manager-78f8948974-7cn4x\" (UID: \"039d67fe-f691-49a7-8e38-4a8dc2a2227e\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.994299 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49wsl\" (UniqueName: \"kubernetes.io/projected/4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4-kube-api-access-49wsl\") pod \"telemetry-operator-controller-manager-76cc84c6bb-c4q9q\" (UID: \"4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" Dec 06 07:12:32 crc kubenswrapper[4945]: I1206 07:12:32.996685 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz8tq\" (UniqueName: \"kubernetes.io/projected/716df590-6583-4b4b-bfdc-4d6bb2ac05a4-kube-api-access-sz8tq\") pod \"swift-operator-controller-manager-5f8c65bbfc-6tvc7\" (UID: \"716df590-6583-4b4b-bfdc-4d6bb2ac05a4\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.008185 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.048172 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t"] Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.049480 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t"] Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.049627 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.063842 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.095544 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.095598 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blfjj\" (UniqueName: \"kubernetes.io/projected/f19785e5-2908-48b0-aa6e-ea4f37bbd506-kube-api-access-blfjj\") pod \"test-operator-controller-manager-5854674fcc-v5krd\" (UID: \"f19785e5-2908-48b0-aa6e-ea4f37bbd506\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.095707 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.095756 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pwnr\" (UniqueName: \"kubernetes.io/projected/c0722d35-7779-4073-851f-0171050b6141-kube-api-access-6pwnr\") pod \"watcher-operator-controller-manager-769dc69bc-2vzqx\" (UID: \"c0722d35-7779-4073-851f-0171050b6141\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.095782 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8bs7\" (UniqueName: \"kubernetes.io/projected/ad774804-402d-45d6-957f-d7ab18bd2f7f-kube-api-access-r8bs7\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.099052 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.099682 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.100108 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-htlv7" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.162303 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.206469 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pwnr\" (UniqueName: \"kubernetes.io/projected/c0722d35-7779-4073-851f-0171050b6141-kube-api-access-6pwnr\") pod \"watcher-operator-controller-manager-769dc69bc-2vzqx\" (UID: \"c0722d35-7779-4073-851f-0171050b6141\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.206520 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8bs7\" (UniqueName: \"kubernetes.io/projected/ad774804-402d-45d6-957f-d7ab18bd2f7f-kube-api-access-r8bs7\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.206611 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.206675 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:33 crc kubenswrapper[4945]: E1206 07:12:33.206847 4945 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 07:12:33 crc kubenswrapper[4945]: E1206 07:12:33.206903 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs podName:ad774804-402d-45d6-957f-d7ab18bd2f7f nodeName:}" failed. No retries permitted until 2025-12-06 07:12:33.706886155 +0000 UTC m=+1167.161747199 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-f6b9t" (UID: "ad774804-402d-45d6-957f-d7ab18bd2f7f") : secret "metrics-server-cert" not found Dec 06 07:12:33 crc kubenswrapper[4945]: E1206 07:12:33.207339 4945 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 07:12:33 crc kubenswrapper[4945]: E1206 07:12:33.207375 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs podName:ad774804-402d-45d6-957f-d7ab18bd2f7f nodeName:}" failed. No retries permitted until 2025-12-06 07:12:33.707363388 +0000 UTC m=+1167.162224432 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-f6b9t" (UID: "ad774804-402d-45d6-957f-d7ab18bd2f7f") : secret "webhook-server-cert" not found Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.255314 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.290994 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8bs7\" (UniqueName: \"kubernetes.io/projected/ad774804-402d-45d6-957f-d7ab18bd2f7f-kube-api-access-r8bs7\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.291249 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pwnr\" (UniqueName: \"kubernetes.io/projected/c0722d35-7779-4073-851f-0171050b6141-kube-api-access-6pwnr\") pod \"watcher-operator-controller-manager-769dc69bc-2vzqx\" (UID: \"c0722d35-7779-4073-851f-0171050b6141\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.346387 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blfjj\" (UniqueName: \"kubernetes.io/projected/f19785e5-2908-48b0-aa6e-ea4f37bbd506-kube-api-access-blfjj\") pod \"test-operator-controller-manager-5854674fcc-v5krd\" (UID: \"f19785e5-2908-48b0-aa6e-ea4f37bbd506\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.381585 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8"] Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.382932 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.388489 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-qk4gl" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.417970 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.422038 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:33 crc kubenswrapper[4945]: E1206 07:12:33.422477 4945 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 07:12:33 crc kubenswrapper[4945]: E1206 07:12:33.422534 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert podName:19942e24-acd1-469c-94b2-37063e88c190 nodeName:}" failed. No retries permitted until 2025-12-06 07:12:34.422516081 +0000 UTC m=+1167.877377125 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f56d49f" (UID: "19942e24-acd1-469c-94b2-37063e88c190") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.443134 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8"] Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.455531 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.497945 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f"] Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.524429 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zmfr\" (UniqueName: \"kubernetes.io/projected/4fe7c3eb-89a4-4935-8e19-935f98f8562b-kube-api-access-9zmfr\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bd5w8\" (UID: \"4fe7c3eb-89a4-4935-8e19-935f98f8562b\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.630646 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zmfr\" (UniqueName: \"kubernetes.io/projected/4fe7c3eb-89a4-4935-8e19-935f98f8562b-kube-api-access-9zmfr\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bd5w8\" (UID: \"4fe7c3eb-89a4-4935-8e19-935f98f8562b\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.668479 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.681245 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zmfr\" (UniqueName: \"kubernetes.io/projected/4fe7c3eb-89a4-4935-8e19-935f98f8562b-kube-api-access-9zmfr\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bd5w8\" (UID: \"4fe7c3eb-89a4-4935-8e19-935f98f8562b\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.731721 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.731894 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:33 crc kubenswrapper[4945]: E1206 07:12:33.732009 4945 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 07:12:33 crc kubenswrapper[4945]: E1206 07:12:33.732073 4945 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 07:12:33 crc kubenswrapper[4945]: E1206 07:12:33.732114 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs podName:ad774804-402d-45d6-957f-d7ab18bd2f7f nodeName:}" failed. No retries permitted until 2025-12-06 07:12:34.732088371 +0000 UTC m=+1168.186949415 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-f6b9t" (UID: "ad774804-402d-45d6-957f-d7ab18bd2f7f") : secret "metrics-server-cert" not found Dec 06 07:12:33 crc kubenswrapper[4945]: E1206 07:12:33.732151 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs podName:ad774804-402d-45d6-957f-d7ab18bd2f7f nodeName:}" failed. No retries permitted until 2025-12-06 07:12:34.732125932 +0000 UTC m=+1168.186987176 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-f6b9t" (UID: "ad774804-402d-45d6-957f-d7ab18bd2f7f") : secret "webhook-server-cert" not found Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.738814 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" event={"ID":"afd67b97-8d07-4522-b37d-019abf73ee85","Type":"ContainerStarted","Data":"4dcc9c379e8225f8df33c566ca78ebf05b9093f7b30dfb50360b653b3a1898dc"} Dec 06 07:12:33 crc kubenswrapper[4945]: I1206 07:12:33.973789 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8" Dec 06 07:12:34 crc kubenswrapper[4945]: I1206 07:12:34.036733 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:34 crc kubenswrapper[4945]: E1206 07:12:34.037532 4945 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 07:12:34 crc kubenswrapper[4945]: E1206 07:12:34.037594 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert podName:db9c0007-43e1-4cb2-ae8a-44a4731254b7 nodeName:}" failed. No retries permitted until 2025-12-06 07:12:36.037573971 +0000 UTC m=+1169.492435015 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert") pod "infra-operator-controller-manager-57548d458d-qhvjz" (UID: "db9c0007-43e1-4cb2-ae8a-44a4731254b7") : secret "infra-operator-webhook-server-cert" not found Dec 06 07:12:34 crc kubenswrapper[4945]: I1206 07:12:34.272936 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4"] Dec 06 07:12:34 crc kubenswrapper[4945]: I1206 07:12:34.443128 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:34 crc kubenswrapper[4945]: E1206 07:12:34.443347 4945 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 07:12:34 crc kubenswrapper[4945]: E1206 07:12:34.443616 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert podName:19942e24-acd1-469c-94b2-37063e88c190 nodeName:}" failed. No retries permitted until 2025-12-06 07:12:36.443596174 +0000 UTC m=+1169.898457218 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f56d49f" (UID: "19942e24-acd1-469c-94b2-37063e88c190") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 07:12:34 crc kubenswrapper[4945]: I1206 07:12:34.580515 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g"] Dec 06 07:12:34 crc kubenswrapper[4945]: I1206 07:12:34.589812 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr"] Dec 06 07:12:34 crc kubenswrapper[4945]: W1206 07:12:34.596954 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d50597e_adf1_4ce6_8f1c_f2acfcbbd0cc.slice/crio-1a430e85299558e4be06815d99c7a28cbe83d07d43add0d87b82e7228f86a513 WatchSource:0}: Error finding container 1a430e85299558e4be06815d99c7a28cbe83d07d43add0d87b82e7228f86a513: Status 404 returned error can't find the container with id 1a430e85299558e4be06815d99c7a28cbe83d07d43add0d87b82e7228f86a513 Dec 06 07:12:34 crc kubenswrapper[4945]: I1206 07:12:34.748786 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" event={"ID":"7305df73-ea7d-43f5-9119-9c12942796bc","Type":"ContainerStarted","Data":"2a97a814a4fc67a0fcd489da30b93a412ab41484ef39dcd414d8a51e48d0997c"} Dec 06 07:12:34 crc kubenswrapper[4945]: I1206 07:12:34.748794 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:34 crc kubenswrapper[4945]: E1206 07:12:34.748984 4945 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 07:12:34 crc kubenswrapper[4945]: E1206 07:12:34.749041 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs podName:ad774804-402d-45d6-957f-d7ab18bd2f7f nodeName:}" failed. No retries permitted until 2025-12-06 07:12:36.749022813 +0000 UTC m=+1170.203883857 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-f6b9t" (UID: "ad774804-402d-45d6-957f-d7ab18bd2f7f") : secret "metrics-server-cert" not found Dec 06 07:12:34 crc kubenswrapper[4945]: I1206 07:12:34.749193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:34 crc kubenswrapper[4945]: E1206 07:12:34.749561 4945 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 07:12:34 crc kubenswrapper[4945]: E1206 07:12:34.749685 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs podName:ad774804-402d-45d6-957f-d7ab18bd2f7f nodeName:}" failed. No retries permitted until 2025-12-06 07:12:36.74963269 +0000 UTC m=+1170.204493794 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-f6b9t" (UID: "ad774804-402d-45d6-957f-d7ab18bd2f7f") : secret "webhook-server-cert" not found Dec 06 07:12:34 crc kubenswrapper[4945]: I1206 07:12:34.750401 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" event={"ID":"31beec86-b14f-43c6-9307-a776454ed9b7","Type":"ContainerStarted","Data":"5570ae4ca0f303f823e5c452f83fc0185fb7f1e5103af985001b9e552cb5d637"} Dec 06 07:12:34 crc kubenswrapper[4945]: I1206 07:12:34.752513 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" event={"ID":"6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc","Type":"ContainerStarted","Data":"1a430e85299558e4be06815d99c7a28cbe83d07d43add0d87b82e7228f86a513"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.063929 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.076964 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8"] Dec 06 07:12:35 crc kubenswrapper[4945]: W1206 07:12:35.091806 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf32f27d3_90de_4c9e_8664_68e149e42ffc.slice/crio-8bce676b53ca1786c446d03cc118a1bedf7a77595f0d91fe50d0b2a726d45562 WatchSource:0}: Error finding container 8bce676b53ca1786c446d03cc118a1bedf7a77595f0d91fe50d0b2a726d45562: Status 404 returned error can't find the container with id 8bce676b53ca1786c446d03cc118a1bedf7a77595f0d91fe50d0b2a726d45562 Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.099783 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.116403 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.140619 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.156748 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.174174 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.203300 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.208403 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.217423 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-v5krd"] Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.219286 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zmgz7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-n8gdh_openstack-operators(8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.224054 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zmgz7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-n8gdh_openstack-operators(8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.225249 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" podUID="8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd" Dec 06 07:12:35 crc kubenswrapper[4945]: W1206 07:12:35.229590 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod716df590_6583_4b4b_bfdc_4d6bb2ac05a4.slice/crio-872020d5ab6d825748ebbe8fef03b30e5750db9ccef58445d7d9435c78174c3c WatchSource:0}: Error finding container 872020d5ab6d825748ebbe8fef03b30e5750db9ccef58445d7d9435c78174c3c: Status 404 returned error can't find the container with id 872020d5ab6d825748ebbe8fef03b30e5750db9ccef58445d7d9435c78174c3c Dec 06 07:12:35 crc kubenswrapper[4945]: W1206 07:12:35.229929 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod039d67fe_f691_49a7_8e38_4a8dc2a2227e.slice/crio-6293f94460bd9d8d31f1e2fb88b2ed2b4b6980309ae42b1c483b196c03b1de0d WatchSource:0}: Error finding container 6293f94460bd9d8d31f1e2fb88b2ed2b4b6980309ae42b1c483b196c03b1de0d: Status 404 returned error can't find the container with id 6293f94460bd9d8d31f1e2fb88b2ed2b4b6980309ae42b1c483b196c03b1de0d Dec 06 07:12:35 crc kubenswrapper[4945]: W1206 07:12:35.231203 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc925c68e_4099_403e_b8ba_84d2d9bc689e.slice/crio-09516eaa92311c4784f8ea46ceaf510bfdb351b1a5c299c3f6d53b07100c5a53 WatchSource:0}: Error finding container 09516eaa92311c4784f8ea46ceaf510bfdb351b1a5c299c3f6d53b07100c5a53: Status 404 returned error can't find the container with id 09516eaa92311c4784f8ea46ceaf510bfdb351b1a5c299c3f6d53b07100c5a53 Dec 06 07:12:35 crc kubenswrapper[4945]: W1206 07:12:35.234391 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bfd7a39_d83d_42de_bb30_0d1dc17bedca.slice/crio-dbc54d89ffb9c22f5897aa45166b8c10848f031fd2a40c9222d971257cee141b WatchSource:0}: Error finding container dbc54d89ffb9c22f5897aa45166b8c10848f031fd2a40c9222d971257cee141b: Status 404 returned error can't find the container with id dbc54d89ffb9c22f5897aa45166b8c10848f031fd2a40c9222d971257cee141b Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.237346 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7"] Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.237527 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sz8tq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6tvc7_openstack-operators(716df590-6583-4b4b-bfdc-4d6bb2ac05a4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.237651 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qr52l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-7cn4x_openstack-operators(039d67fe-f691-49a7-8e38-4a8dc2a2227e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.239572 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hfpxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-pd42f_openstack-operators(9bfd7a39-d83d-42de-bb30-0d1dc17bedca): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.240006 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qr52l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-7cn4x_openstack-operators(039d67fe-f691-49a7-8e38-4a8dc2a2227e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.240812 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sz8tq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6tvc7_openstack-operators(716df590-6583-4b4b-bfdc-4d6bb2ac05a4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.241148 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" podUID="039d67fe-f691-49a7-8e38-4a8dc2a2227e" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.242123 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" podUID="716df590-6583-4b4b-bfdc-4d6bb2ac05a4" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.253697 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-smt2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-cgbvh_openstack-operators(c925c68e-4099-403e-b8ba-84d2d9bc689e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.253994 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hfpxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-pd42f_openstack-operators(9bfd7a39-d83d-42de-bb30-0d1dc17bedca): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.254105 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6pwnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-2vzqx_openstack-operators(c0722d35-7779-4073-851f-0171050b6141): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.255178 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" podUID="9bfd7a39-d83d-42de-bb30-0d1dc17bedca" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.257205 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6pwnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-2vzqx_openstack-operators(c0722d35-7779-4073-851f-0171050b6141): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.258647 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" podUID="c0722d35-7779-4073-851f-0171050b6141" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.260045 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-smt2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-cgbvh_openstack-operators(c925c68e-4099-403e-b8ba-84d2d9bc689e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.261953 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" podUID="c925c68e-4099-403e-b8ba-84d2d9bc689e" Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.270918 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-pd42f"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.296116 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.312402 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.320140 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.327162 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx"] Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.768060 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" event={"ID":"039d67fe-f691-49a7-8e38-4a8dc2a2227e","Type":"ContainerStarted","Data":"6293f94460bd9d8d31f1e2fb88b2ed2b4b6980309ae42b1c483b196c03b1de0d"} Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.772118 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" podUID="039d67fe-f691-49a7-8e38-4a8dc2a2227e" Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.773168 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" event={"ID":"4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4","Type":"ContainerStarted","Data":"561c936e60a75815612ffb76e0bda710c2642bb9b39bbabdb8ec044277545f3f"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.774901 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" event={"ID":"f32f27d3-90de-4c9e-8664-68e149e42ffc","Type":"ContainerStarted","Data":"8bce676b53ca1786c446d03cc118a1bedf7a77595f0d91fe50d0b2a726d45562"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.784383 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" event={"ID":"3c4830bb-bc9c-447e-bb30-9c481fae8234","Type":"ContainerStarted","Data":"7496ecd067d59dd7443f12e5f1a0557209c91f1f9d9d41dde735677b8e9c3910"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.788778 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" event={"ID":"c925c68e-4099-403e-b8ba-84d2d9bc689e","Type":"ContainerStarted","Data":"09516eaa92311c4784f8ea46ceaf510bfdb351b1a5c299c3f6d53b07100c5a53"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.790667 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" event={"ID":"8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd","Type":"ContainerStarted","Data":"19f3473ca4301201eb97462ca705a4364c42dd32589abd719c4b0e2cff5bafe6"} Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.791043 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" podUID="c925c68e-4099-403e-b8ba-84d2d9bc689e" Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.792840 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" event={"ID":"c0722d35-7779-4073-851f-0171050b6141","Type":"ContainerStarted","Data":"b8546bb56c38a27916065c6d14a6c9181c0d40a67ee0e68ea68a644165c417e0"} Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.793520 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" podUID="8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd" Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.794830 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" podUID="c0722d35-7779-4073-851f-0171050b6141" Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.794989 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" event={"ID":"9bfd7a39-d83d-42de-bb30-0d1dc17bedca","Type":"ContainerStarted","Data":"dbc54d89ffb9c22f5897aa45166b8c10848f031fd2a40c9222d971257cee141b"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.796178 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" event={"ID":"7c3f47b7-f1f2-408d-80d0-03a0ea9286ea","Type":"ContainerStarted","Data":"fe5ae3ecdc037239c2e9f07254a3a50cd284018dd92c656272d6c05853685279"} Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.796368 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" podUID="9bfd7a39-d83d-42de-bb30-0d1dc17bedca" Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.797495 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" event={"ID":"f19785e5-2908-48b0-aa6e-ea4f37bbd506","Type":"ContainerStarted","Data":"7fde911fded7079dd569ba6ff7ff627f2562603138ebb2d80bad3de8632361d5"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.799537 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" event={"ID":"716df590-6583-4b4b-bfdc-4d6bb2ac05a4","Type":"ContainerStarted","Data":"872020d5ab6d825748ebbe8fef03b30e5750db9ccef58445d7d9435c78174c3c"} Dec 06 07:12:35 crc kubenswrapper[4945]: E1206 07:12:35.802766 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" podUID="716df590-6583-4b4b-bfdc-4d6bb2ac05a4" Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.804097 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" event={"ID":"6c628f21-0f62-4761-9dd1-f6377a5633a5","Type":"ContainerStarted","Data":"ddbcbb477c4d8e128579477f63cf4c6f923a59f5ca367e789a48356625c3fc55"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.809261 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" event={"ID":"0e92db58-a1fc-4035-b635-f274d7411813","Type":"ContainerStarted","Data":"a947f69cd82cadfb9f56a6d7d9e2b49dca671aeea57369e252102d52b4d7fb5d"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.813405 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8" event={"ID":"4fe7c3eb-89a4-4935-8e19-935f98f8562b","Type":"ContainerStarted","Data":"37c39316736494451efad453859a1bd0d31e4a206a0030c335c6a5ca6a41f0a1"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.814792 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" event={"ID":"40318b4d-ed4f-478b-8333-8e1cb0729b34","Type":"ContainerStarted","Data":"4e976f8cbd8c2b4357ed44a8efa23e846ca8e2993885824988b6f0ed77fe05cc"} Dec 06 07:12:35 crc kubenswrapper[4945]: I1206 07:12:35.816448 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" event={"ID":"b89bffc9-f64a-4d0b-b5e5-104db0d681dd","Type":"ContainerStarted","Data":"02573244c2ff8e3aa74df6645772b85d4e978c8b76705bbc6d2fcad6e82b4946"} Dec 06 07:12:36 crc kubenswrapper[4945]: I1206 07:12:36.088051 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.088300 4945 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.088391 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert podName:db9c0007-43e1-4cb2-ae8a-44a4731254b7 nodeName:}" failed. No retries permitted until 2025-12-06 07:12:40.088365341 +0000 UTC m=+1173.543226385 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert") pod "infra-operator-controller-manager-57548d458d-qhvjz" (UID: "db9c0007-43e1-4cb2-ae8a-44a4731254b7") : secret "infra-operator-webhook-server-cert" not found Dec 06 07:12:36 crc kubenswrapper[4945]: I1206 07:12:36.498273 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.498396 4945 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.498459 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert podName:19942e24-acd1-469c-94b2-37063e88c190 nodeName:}" failed. No retries permitted until 2025-12-06 07:12:40.498442554 +0000 UTC m=+1173.953303598 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f56d49f" (UID: "19942e24-acd1-469c-94b2-37063e88c190") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 07:12:36 crc kubenswrapper[4945]: I1206 07:12:36.804868 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:36 crc kubenswrapper[4945]: I1206 07:12:36.804981 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.805142 4945 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.805191 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs podName:ad774804-402d-45d6-957f-d7ab18bd2f7f nodeName:}" failed. No retries permitted until 2025-12-06 07:12:40.805176558 +0000 UTC m=+1174.260037602 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-f6b9t" (UID: "ad774804-402d-45d6-957f-d7ab18bd2f7f") : secret "metrics-server-cert" not found Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.805572 4945 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.805608 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs podName:ad774804-402d-45d6-957f-d7ab18bd2f7f nodeName:}" failed. No retries permitted until 2025-12-06 07:12:40.805597859 +0000 UTC m=+1174.260458903 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-f6b9t" (UID: "ad774804-402d-45d6-957f-d7ab18bd2f7f") : secret "webhook-server-cert" not found Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.829105 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" podUID="c0722d35-7779-4073-851f-0171050b6141" Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.829191 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" podUID="c925c68e-4099-403e-b8ba-84d2d9bc689e" Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.829741 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" podUID="039d67fe-f691-49a7-8e38-4a8dc2a2227e" Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.829806 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" podUID="716df590-6583-4b4b-bfdc-4d6bb2ac05a4" Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.829864 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" podUID="9bfd7a39-d83d-42de-bb30-0d1dc17bedca" Dec 06 07:12:36 crc kubenswrapper[4945]: E1206 07:12:36.829918 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" podUID="8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd" Dec 06 07:12:40 crc kubenswrapper[4945]: I1206 07:12:40.220727 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:40 crc kubenswrapper[4945]: E1206 07:12:40.221212 4945 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 07:12:40 crc kubenswrapper[4945]: E1206 07:12:40.221271 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert podName:db9c0007-43e1-4cb2-ae8a-44a4731254b7 nodeName:}" failed. No retries permitted until 2025-12-06 07:12:48.221252502 +0000 UTC m=+1181.676113546 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert") pod "infra-operator-controller-manager-57548d458d-qhvjz" (UID: "db9c0007-43e1-4cb2-ae8a-44a4731254b7") : secret "infra-operator-webhook-server-cert" not found Dec 06 07:12:40 crc kubenswrapper[4945]: I1206 07:12:40.576162 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:40 crc kubenswrapper[4945]: E1206 07:12:40.576377 4945 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 07:12:40 crc kubenswrapper[4945]: E1206 07:12:40.576456 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert podName:19942e24-acd1-469c-94b2-37063e88c190 nodeName:}" failed. No retries permitted until 2025-12-06 07:12:48.576436369 +0000 UTC m=+1182.031297413 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f56d49f" (UID: "19942e24-acd1-469c-94b2-37063e88c190") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 07:12:40 crc kubenswrapper[4945]: I1206 07:12:40.783440 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-269rv" podUID="58b9a31e-1915-40f7-8e45-cca0242826aa" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:12:40 crc kubenswrapper[4945]: I1206 07:12:40.880112 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:40 crc kubenswrapper[4945]: E1206 07:12:40.880416 4945 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 07:12:40 crc kubenswrapper[4945]: I1206 07:12:40.880589 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:40 crc kubenswrapper[4945]: E1206 07:12:40.880704 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs podName:ad774804-402d-45d6-957f-d7ab18bd2f7f nodeName:}" failed. No retries permitted until 2025-12-06 07:12:48.880660365 +0000 UTC m=+1182.335521399 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-f6b9t" (UID: "ad774804-402d-45d6-957f-d7ab18bd2f7f") : secret "metrics-server-cert" not found Dec 06 07:12:40 crc kubenswrapper[4945]: E1206 07:12:40.880766 4945 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 07:12:40 crc kubenswrapper[4945]: E1206 07:12:40.880996 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs podName:ad774804-402d-45d6-957f-d7ab18bd2f7f nodeName:}" failed. No retries permitted until 2025-12-06 07:12:48.880973203 +0000 UTC m=+1182.335834317 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-f6b9t" (UID: "ad774804-402d-45d6-957f-d7ab18bd2f7f") : secret "webhook-server-cert" not found Dec 06 07:12:47 crc kubenswrapper[4945]: E1206 07:12:47.869767 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 06 07:12:47 crc kubenswrapper[4945]: E1206 07:12:47.870595 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jnx45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-lwkxm_openstack-operators(6c628f21-0f62-4761-9dd1-f6377a5633a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:12:48 crc kubenswrapper[4945]: I1206 07:12:48.239146 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:48 crc kubenswrapper[4945]: I1206 07:12:48.244752 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db9c0007-43e1-4cb2-ae8a-44a4731254b7-cert\") pod \"infra-operator-controller-manager-57548d458d-qhvjz\" (UID: \"db9c0007-43e1-4cb2-ae8a-44a4731254b7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:48 crc kubenswrapper[4945]: I1206 07:12:48.294638 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:12:48 crc kubenswrapper[4945]: I1206 07:12:48.646511 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:48 crc kubenswrapper[4945]: I1206 07:12:48.652531 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19942e24-acd1-469c-94b2-37063e88c190-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56d49f\" (UID: \"19942e24-acd1-469c-94b2-37063e88c190\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:48 crc kubenswrapper[4945]: I1206 07:12:48.927896 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:12:48 crc kubenswrapper[4945]: I1206 07:12:48.951060 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:48 crc kubenswrapper[4945]: I1206 07:12:48.951202 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:48 crc kubenswrapper[4945]: I1206 07:12:48.954870 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:48 crc kubenswrapper[4945]: I1206 07:12:48.955471 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad774804-402d-45d6-957f-d7ab18bd2f7f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-f6b9t\" (UID: \"ad774804-402d-45d6-957f-d7ab18bd2f7f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:49 crc kubenswrapper[4945]: I1206 07:12:49.051478 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:12:51 crc kubenswrapper[4945]: E1206 07:12:51.677471 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 06 07:12:51 crc kubenswrapper[4945]: E1206 07:12:51.678001 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wf2jv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-r7zm6_openstack-operators(b89bffc9-f64a-4d0b-b5e5-104db0d681dd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:12:53 crc kubenswrapper[4945]: E1206 07:12:53.008713 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 06 07:12:53 crc kubenswrapper[4945]: E1206 07:12:53.009254 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qhb4h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-kjxk4_openstack-operators(31beec86-b14f-43c6-9307-a776454ed9b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:12:54 crc kubenswrapper[4945]: E1206 07:12:54.159072 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 06 07:12:54 crc kubenswrapper[4945]: E1206 07:12:54.159320 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s7srq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-wqq5g_openstack-operators(7305df73-ea7d-43f5-9119-9c12942796bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:12:57 crc kubenswrapper[4945]: E1206 07:12:57.235468 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 06 07:12:57 crc kubenswrapper[4945]: E1206 07:12:57.235973 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l4g2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-cchx8_openstack-operators(f32f27d3-90de-4c9e-8664-68e149e42ffc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:12:57 crc kubenswrapper[4945]: E1206 07:12:57.748604 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 06 07:12:57 crc kubenswrapper[4945]: E1206 07:12:57.748834 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-blfjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-v5krd_openstack-operators(f19785e5-2908-48b0-aa6e-ea4f37bbd506): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:00 crc kubenswrapper[4945]: E1206 07:13:00.295515 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 06 07:13:00 crc kubenswrapper[4945]: E1206 07:13:00.296020 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tkmrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-bpsxr_openstack-operators(6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:00 crc kubenswrapper[4945]: E1206 07:13:00.758715 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 06 07:13:00 crc kubenswrapper[4945]: E1206 07:13:00.758912 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bgsw2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-5njm2_openstack-operators(0e92db58-a1fc-4035-b635-f274d7411813): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:02 crc kubenswrapper[4945]: E1206 07:13:02.397476 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 06 07:13:02 crc kubenswrapper[4945]: E1206 07:13:02.398350 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x54sz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-x98dd_openstack-operators(7c3f47b7-f1f2-408d-80d0-03a0ea9286ea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:03 crc kubenswrapper[4945]: E1206 07:13:03.232499 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Dec 06 07:13:03 crc kubenswrapper[4945]: E1206 07:13:03.232742 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-49wsl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-c4q9q_openstack-operators(4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:04 crc kubenswrapper[4945]: E1206 07:13:04.287949 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 06 07:13:04 crc kubenswrapper[4945]: E1206 07:13:04.288157 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sz8tq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6tvc7_openstack-operators(716df590-6583-4b4b-bfdc-4d6bb2ac05a4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:06 crc kubenswrapper[4945]: E1206 07:13:06.721056 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 06 07:13:06 crc kubenswrapper[4945]: E1206 07:13:06.721613 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-smt2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-cgbvh_openstack-operators(c925c68e-4099-403e-b8ba-84d2d9bc689e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:07 crc kubenswrapper[4945]: E1206 07:13:07.394624 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 06 07:13:07 crc kubenswrapper[4945]: E1206 07:13:07.395257 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qr52l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-7cn4x_openstack-operators(039d67fe-f691-49a7-8e38-4a8dc2a2227e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:08 crc kubenswrapper[4945]: E1206 07:13:08.072199 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 06 07:13:08 crc kubenswrapper[4945]: E1206 07:13:08.073016 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6pwnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-2vzqx_openstack-operators(c0722d35-7779-4073-851f-0171050b6141): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:08 crc kubenswrapper[4945]: E1206 07:13:08.737057 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 06 07:13:08 crc kubenswrapper[4945]: E1206 07:13:08.737260 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hfpxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-pd42f_openstack-operators(9bfd7a39-d83d-42de-bb30-0d1dc17bedca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:08 crc kubenswrapper[4945]: I1206 07:13:08.842490 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:13:08 crc kubenswrapper[4945]: I1206 07:13:08.842611 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:13:09 crc kubenswrapper[4945]: E1206 07:13:09.397358 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 06 07:13:09 crc kubenswrapper[4945]: E1206 07:13:09.397708 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zmgz7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-n8gdh_openstack-operators(8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:10 crc kubenswrapper[4945]: E1206 07:13:10.579067 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 06 07:13:10 crc kubenswrapper[4945]: E1206 07:13:10.579652 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9zmfr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-bd5w8_openstack-operators(4fe7c3eb-89a4-4935-8e19-935f98f8562b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:13:10 crc kubenswrapper[4945]: E1206 07:13:10.580839 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8" podUID="4fe7c3eb-89a4-4935-8e19-935f98f8562b" Dec 06 07:13:11 crc kubenswrapper[4945]: I1206 07:13:11.250451 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f"] Dec 06 07:13:11 crc kubenswrapper[4945]: I1206 07:13:11.297006 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz"] Dec 06 07:13:11 crc kubenswrapper[4945]: I1206 07:13:11.307179 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t"] Dec 06 07:13:11 crc kubenswrapper[4945]: W1206 07:13:11.356193 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19942e24_acd1_469c_94b2_37063e88c190.slice/crio-beaa5dbe3b3bda74885d3241d652cd208079a82d8751654271fb4a91cdadabe0 WatchSource:0}: Error finding container beaa5dbe3b3bda74885d3241d652cd208079a82d8751654271fb4a91cdadabe0: Status 404 returned error can't find the container with id beaa5dbe3b3bda74885d3241d652cd208079a82d8751654271fb4a91cdadabe0 Dec 06 07:13:11 crc kubenswrapper[4945]: W1206 07:13:11.359178 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb9c0007_43e1_4cb2_ae8a_44a4731254b7.slice/crio-fb0d8acc1550f77684edfc785a5ddd44fda3f816e2d8dd58f6c5d5a8671b1ee7 WatchSource:0}: Error finding container fb0d8acc1550f77684edfc785a5ddd44fda3f816e2d8dd58f6c5d5a8671b1ee7: Status 404 returned error can't find the container with id fb0d8acc1550f77684edfc785a5ddd44fda3f816e2d8dd58f6c5d5a8671b1ee7 Dec 06 07:13:11 crc kubenswrapper[4945]: W1206 07:13:11.370522 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad774804_402d_45d6_957f_d7ab18bd2f7f.slice/crio-56ba377cdd1d43ececd59522ea9f943f781493db89f525e3d8155fafe58baa68 WatchSource:0}: Error finding container 56ba377cdd1d43ececd59522ea9f943f781493db89f525e3d8155fafe58baa68: Status 404 returned error can't find the container with id 56ba377cdd1d43ececd59522ea9f943f781493db89f525e3d8155fafe58baa68 Dec 06 07:13:11 crc kubenswrapper[4945]: I1206 07:13:11.451722 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" event={"ID":"19942e24-acd1-469c-94b2-37063e88c190","Type":"ContainerStarted","Data":"beaa5dbe3b3bda74885d3241d652cd208079a82d8751654271fb4a91cdadabe0"} Dec 06 07:13:11 crc kubenswrapper[4945]: I1206 07:13:11.452966 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" event={"ID":"ad774804-402d-45d6-957f-d7ab18bd2f7f","Type":"ContainerStarted","Data":"56ba377cdd1d43ececd59522ea9f943f781493db89f525e3d8155fafe58baa68"} Dec 06 07:13:11 crc kubenswrapper[4945]: I1206 07:13:11.454534 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" event={"ID":"db9c0007-43e1-4cb2-ae8a-44a4731254b7","Type":"ContainerStarted","Data":"fb0d8acc1550f77684edfc785a5ddd44fda3f816e2d8dd58f6c5d5a8671b1ee7"} Dec 06 07:13:11 crc kubenswrapper[4945]: E1206 07:13:11.456003 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8" podUID="4fe7c3eb-89a4-4935-8e19-935f98f8562b" Dec 06 07:13:12 crc kubenswrapper[4945]: I1206 07:13:12.475028 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" event={"ID":"afd67b97-8d07-4522-b37d-019abf73ee85","Type":"ContainerStarted","Data":"2294f9c8a71fd746f378415b19c18bbb2f516f1659d2c966b0ac6e00ad69c0d6"} Dec 06 07:13:12 crc kubenswrapper[4945]: I1206 07:13:12.477688 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" event={"ID":"3c4830bb-bc9c-447e-bb30-9c481fae8234","Type":"ContainerStarted","Data":"e7ba8315f50025a133b540b13b22a98de33544349be3b51c36add994a4c5d62b"} Dec 06 07:13:15 crc kubenswrapper[4945]: I1206 07:13:15.624966 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" event={"ID":"ad774804-402d-45d6-957f-d7ab18bd2f7f","Type":"ContainerStarted","Data":"ed807d7172ce955b9897f9f83838be1abc99fc6f76da0ac07694a06da7f3d3e9"} Dec 06 07:13:15 crc kubenswrapper[4945]: I1206 07:13:15.625829 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:13:15 crc kubenswrapper[4945]: I1206 07:13:15.628521 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" event={"ID":"40318b4d-ed4f-478b-8333-8e1cb0729b34","Type":"ContainerStarted","Data":"155359a9afe762489f2d179b291c9bcdc0e8707c3da5ee7979f2bff466826854"} Dec 06 07:13:15 crc kubenswrapper[4945]: I1206 07:13:15.666716 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" podStartSLOduration=43.666699419 podStartE2EDuration="43.666699419s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:13:15.661496319 +0000 UTC m=+1209.116357363" watchObservedRunningTime="2025-12-06 07:13:15.666699419 +0000 UTC m=+1209.121560463" Dec 06 07:13:16 crc kubenswrapper[4945]: E1206 07:13:16.937437 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" podUID="6c628f21-0f62-4761-9dd1-f6377a5633a5" Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.136079 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" podUID="f19785e5-2908-48b0-aa6e-ea4f37bbd506" Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.235944 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" podUID="31beec86-b14f-43c6-9307-a776454ed9b7" Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.295002 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" podUID="c925c68e-4099-403e-b8ba-84d2d9bc689e" Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.306463 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" podUID="f32f27d3-90de-4c9e-8664-68e149e42ffc" Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.343645 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" podUID="039d67fe-f691-49a7-8e38-4a8dc2a2227e" Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.363105 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" podUID="b89bffc9-f64a-4d0b-b5e5-104db0d681dd" Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.417781 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" podUID="716df590-6583-4b4b-bfdc-4d6bb2ac05a4" Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.517782 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" podUID="4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4" Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.659268 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" event={"ID":"b89bffc9-f64a-4d0b-b5e5-104db0d681dd","Type":"ContainerStarted","Data":"fcf05e6ae61b05c5ee8199aa70bd5271e697bb425a7a9f9bb96ce4f21aba0852"} Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.672951 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" event={"ID":"f32f27d3-90de-4c9e-8664-68e149e42ffc","Type":"ContainerStarted","Data":"9cbab74d21baefd12b1ea43868d4cf10739bd5a040659bd195870f125ad23e79"} Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.678301 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" event={"ID":"f19785e5-2908-48b0-aa6e-ea4f37bbd506","Type":"ContainerStarted","Data":"ebc8947ceae03ff905f925935128576d7c8b6c92bf03db11e4866778e4b3d47c"} Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.688700 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" event={"ID":"c925c68e-4099-403e-b8ba-84d2d9bc689e","Type":"ContainerStarted","Data":"8a7297a1212f9c0a36fd576528696af12ba16d01d49bc919662308dfed6409a1"} Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.693344 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809\\\"\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" podUID="c925c68e-4099-403e-b8ba-84d2d9bc689e" Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.696074 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" event={"ID":"716df590-6583-4b4b-bfdc-4d6bb2ac05a4","Type":"ContainerStarted","Data":"f5a5deac51291f17dd06d91ee705f85984d0aa9b977b0659822bbe23b27b72cc"} Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.697726 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" podUID="716df590-6583-4b4b-bfdc-4d6bb2ac05a4" Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.711650 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" event={"ID":"4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4","Type":"ContainerStarted","Data":"5d2d6a12acaeada73194282f3b0982133bac8943c54c621eadebd2936b1a40e3"} Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.764612 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" event={"ID":"afd67b97-8d07-4522-b37d-019abf73ee85","Type":"ContainerStarted","Data":"48101f5e3ca155454751fab5ba7eb45d94d7d35ea7034bd1ad954088ea542e55"} Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.767127 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.771533 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.787775 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" event={"ID":"31beec86-b14f-43c6-9307-a776454ed9b7","Type":"ContainerStarted","Data":"65c44aee92063df6636e2f02286b83c6743006d41ee197e35ed2111049e560dc"} Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.807152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" event={"ID":"3c4830bb-bc9c-447e-bb30-9c481fae8234","Type":"ContainerStarted","Data":"f221b21837d0a551bfb9f487938ac8e132fdec6a44828f409c32b9e7bc293eff"} Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.808022 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.870882 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.881631 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" event={"ID":"db9c0007-43e1-4cb2-ae8a-44a4731254b7","Type":"ContainerStarted","Data":"6d77ed4e78836a7d0b329a3b39cd08b3d045126041963a8c290c75e444d9a342"} Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.902159 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" event={"ID":"039d67fe-f691-49a7-8e38-4a8dc2a2227e","Type":"ContainerStarted","Data":"96cacfeaf6ae2eb189cccdbf511f200fa6dbed67fbc5968344576df014617861"} Dec 06 07:13:17 crc kubenswrapper[4945]: E1206 07:13:17.904690 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" podUID="039d67fe-f691-49a7-8e38-4a8dc2a2227e" Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.911188 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" event={"ID":"6c628f21-0f62-4761-9dd1-f6377a5633a5","Type":"ContainerStarted","Data":"68b21bac60fdce9a924383ca788818ab1f38443ba61212f61c9f946190f279b7"} Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.938554 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" event={"ID":"40318b4d-ed4f-478b-8333-8e1cb0729b34","Type":"ContainerStarted","Data":"65435a32f178b0e09400801e668b0bb024490a5dcb1c747ae0402e7960f47acd"} Dec 06 07:13:17 crc kubenswrapper[4945]: I1206 07:13:17.939827 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" Dec 06 07:13:18 crc kubenswrapper[4945]: I1206 07:13:18.086054 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-qd55f" podStartSLOduration=4.252431973 podStartE2EDuration="47.086034793s" podCreationTimestamp="2025-12-06 07:12:31 +0000 UTC" firstStartedPulling="2025-12-06 07:12:33.668178304 +0000 UTC m=+1167.123039348" lastFinishedPulling="2025-12-06 07:13:16.501781124 +0000 UTC m=+1209.956642168" observedRunningTime="2025-12-06 07:13:18.0818397 +0000 UTC m=+1211.536700744" watchObservedRunningTime="2025-12-06 07:13:18.086034793 +0000 UTC m=+1211.540895847" Dec 06 07:13:18 crc kubenswrapper[4945]: I1206 07:13:18.188498 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f6rz7" podStartSLOduration=4.84883318 podStartE2EDuration="46.188462946s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.149270701 +0000 UTC m=+1168.604131745" lastFinishedPulling="2025-12-06 07:13:16.488900467 +0000 UTC m=+1209.943761511" observedRunningTime="2025-12-06 07:13:18.164410749 +0000 UTC m=+1211.619271793" watchObservedRunningTime="2025-12-06 07:13:18.188462946 +0000 UTC m=+1211.643323990" Dec 06 07:13:18 crc kubenswrapper[4945]: I1206 07:13:18.272895 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" podStartSLOduration=4.947408448 podStartE2EDuration="46.272861474s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.164159711 +0000 UTC m=+1168.619020755" lastFinishedPulling="2025-12-06 07:13:16.489612737 +0000 UTC m=+1209.944473781" observedRunningTime="2025-12-06 07:13:18.239611541 +0000 UTC m=+1211.694472575" watchObservedRunningTime="2025-12-06 07:13:18.272861474 +0000 UTC m=+1211.727722518" Dec 06 07:13:20 crc kubenswrapper[4945]: I1206 07:13:20.084370 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-p2gzt" Dec 06 07:13:20 crc kubenswrapper[4945]: E1206 07:13:20.505518 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" podUID="8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd" Dec 06 07:13:20 crc kubenswrapper[4945]: E1206 07:13:20.951729 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" podUID="6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc" Dec 06 07:13:21 crc kubenswrapper[4945]: E1206 07:13:21.017655 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" podUID="c0722d35-7779-4073-851f-0171050b6141" Dec 06 07:13:21 crc kubenswrapper[4945]: E1206 07:13:21.022424 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" podUID="7305df73-ea7d-43f5-9119-9c12942796bc" Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.095539 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" event={"ID":"6c628f21-0f62-4761-9dd1-f6377a5633a5","Type":"ContainerStarted","Data":"c2eb17093c8bc48ebb3e7eb973429f2da8e0133c16235eb354112f3cb01d3774"} Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.095868 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.098388 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" event={"ID":"b89bffc9-f64a-4d0b-b5e5-104db0d681dd","Type":"ContainerStarted","Data":"d2eb7dad5442bbac2a05edea58877a832a56ea480f867dd5e018261e9c689e99"} Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.099372 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.102040 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" event={"ID":"7305df73-ea7d-43f5-9119-9c12942796bc","Type":"ContainerStarted","Data":"32f5bfacde5c198545d190518d92b90f6acd15c6b508d2130e631bdadf49f2be"} Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.105864 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.110501 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" event={"ID":"db9c0007-43e1-4cb2-ae8a-44a4731254b7","Type":"ContainerStarted","Data":"b126538fcf9dcce0fb46f91b5c8e9715be3c07dea59d18c920f559b942877ede"} Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.115477 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" event={"ID":"8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd","Type":"ContainerStarted","Data":"0ef92d8298843e5f4fe7d12a062bea2d6d850fb92d5af66a7b790679107ccbab"} Dec 06 07:13:21 crc kubenswrapper[4945]: E1206 07:13:21.117716 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" podUID="8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd" Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.118846 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" event={"ID":"c0722d35-7779-4073-851f-0171050b6141","Type":"ContainerStarted","Data":"9da83ca9b9825a574ee91d877079b902e45b7c69ce750322d5c5fc49ec79478f"} Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.119129 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" podStartSLOduration=3.623350942 podStartE2EDuration="49.119115573s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.138310506 +0000 UTC m=+1168.593171550" lastFinishedPulling="2025-12-06 07:13:20.634075137 +0000 UTC m=+1214.088936181" observedRunningTime="2025-12-06 07:13:21.113944964 +0000 UTC m=+1214.568806008" watchObservedRunningTime="2025-12-06 07:13:21.119115573 +0000 UTC m=+1214.573976617" Dec 06 07:13:21 crc kubenswrapper[4945]: E1206 07:13:21.120297 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" podUID="c0722d35-7779-4073-851f-0171050b6141" Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.121629 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" event={"ID":"6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc","Type":"ContainerStarted","Data":"80a737d6aa6c8e3abc557680f118cd086ba38b5bcd52fc14c3f9afa7e4d1a0b5"} Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.134816 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" podStartSLOduration=44.002429874 podStartE2EDuration="49.134797485s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:13:11.367139401 +0000 UTC m=+1204.822000445" lastFinishedPulling="2025-12-06 07:13:16.499507012 +0000 UTC m=+1209.954368056" observedRunningTime="2025-12-06 07:13:21.13237611 +0000 UTC m=+1214.587237164" watchObservedRunningTime="2025-12-06 07:13:21.134797485 +0000 UTC m=+1214.589658529" Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.308555 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" podStartSLOduration=3.908599589 podStartE2EDuration="49.308536744s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.217540006 +0000 UTC m=+1168.672401050" lastFinishedPulling="2025-12-06 07:13:20.617477161 +0000 UTC m=+1214.072338205" observedRunningTime="2025-12-06 07:13:21.184508721 +0000 UTC m=+1214.639369765" watchObservedRunningTime="2025-12-06 07:13:21.308536744 +0000 UTC m=+1214.763397788" Dec 06 07:13:21 crc kubenswrapper[4945]: I1206 07:13:21.309599 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" podStartSLOduration=3.890062102 podStartE2EDuration="49.309588433s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.208159864 +0000 UTC m=+1168.663020908" lastFinishedPulling="2025-12-06 07:13:20.627686205 +0000 UTC m=+1214.082547239" observedRunningTime="2025-12-06 07:13:21.293016777 +0000 UTC m=+1214.747877841" watchObservedRunningTime="2025-12-06 07:13:21.309588433 +0000 UTC m=+1214.764449477" Dec 06 07:13:21 crc kubenswrapper[4945]: E1206 07:13:21.520519 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" podUID="0e92db58-a1fc-4035-b635-f274d7411813" Dec 06 07:13:21 crc kubenswrapper[4945]: E1206 07:13:21.527493 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" podUID="7c3f47b7-f1f2-408d-80d0-03a0ea9286ea" Dec 06 07:13:21 crc kubenswrapper[4945]: E1206 07:13:21.545730 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" podUID="9bfd7a39-d83d-42de-bb30-0d1dc17bedca" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.181110 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" event={"ID":"31beec86-b14f-43c6-9307-a776454ed9b7","Type":"ContainerStarted","Data":"f6c942e41dd9856f94cf6486b5e1f9ce4bc60eb7d4c7a9c00dd99135030b9f73"} Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.182129 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.187833 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" event={"ID":"f19785e5-2908-48b0-aa6e-ea4f37bbd506","Type":"ContainerStarted","Data":"482bc08edbfae4787d66c0ae7bdd7d800288d55cdd06405fb6da11e358a7fc4d"} Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.191368 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" event={"ID":"4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4","Type":"ContainerStarted","Data":"eba0fed598b253b2361e3c2693777f35b794d9adc4d50a044f63ebdc1ba4fce5"} Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.191521 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.192963 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" event={"ID":"0e92db58-a1fc-4035-b635-f274d7411813","Type":"ContainerStarted","Data":"b91aca6ceb17d0635308400829c8d9f6a012d63d9c7b00a3b7f2d794812a2c04"} Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.195380 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" event={"ID":"9bfd7a39-d83d-42de-bb30-0d1dc17bedca","Type":"ContainerStarted","Data":"3abb4091b8ede55fe5f6bc0a196fefceaa00a8abb0f6d25f2915a25e540fe52c"} Dec 06 07:13:22 crc kubenswrapper[4945]: E1206 07:13:22.197722 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" podUID="9bfd7a39-d83d-42de-bb30-0d1dc17bedca" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.208775 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" event={"ID":"f32f27d3-90de-4c9e-8664-68e149e42ffc","Type":"ContainerStarted","Data":"dd663e0594920e8fae0b900577e20a57cf2e4c5a22b4f144fce04f40e251002a"} Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.208936 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.209426 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" podStartSLOduration=4.952884079 podStartE2EDuration="51.209396867s" podCreationTimestamp="2025-12-06 07:12:31 +0000 UTC" firstStartedPulling="2025-12-06 07:12:34.392513611 +0000 UTC m=+1167.847374655" lastFinishedPulling="2025-12-06 07:13:20.649026399 +0000 UTC m=+1214.103887443" observedRunningTime="2025-12-06 07:13:22.201910446 +0000 UTC m=+1215.656771500" watchObservedRunningTime="2025-12-06 07:13:22.209396867 +0000 UTC m=+1215.664257921" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.211188 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" event={"ID":"7c3f47b7-f1f2-408d-80d0-03a0ea9286ea","Type":"ContainerStarted","Data":"5dd0f871b458af2387b4a98c1c51385b5429b43e22da46261f6df7985986d158"} Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.214602 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" event={"ID":"19942e24-acd1-469c-94b2-37063e88c190","Type":"ContainerStarted","Data":"6f052e64c4e23798a9d6dc235a259805396b05710e61dbe5fbb93373d9732e1c"} Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.214667 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" event={"ID":"19942e24-acd1-469c-94b2-37063e88c190","Type":"ContainerStarted","Data":"138813ecd6dbce4cd37f2f9a53bd218369ec064cc47ca5722782b397a25acdb5"} Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.214813 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.214867 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.223058 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhvjz" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.306257 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" podStartSLOduration=4.824771504 podStartE2EDuration="50.30622964s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.166995107 +0000 UTC m=+1168.621856151" lastFinishedPulling="2025-12-06 07:13:20.648453243 +0000 UTC m=+1214.103314287" observedRunningTime="2025-12-06 07:13:22.299320254 +0000 UTC m=+1215.754181318" watchObservedRunningTime="2025-12-06 07:13:22.30622964 +0000 UTC m=+1215.761090684" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.358020 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" podStartSLOduration=41.176615305 podStartE2EDuration="50.358003521s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:13:11.36225129 +0000 UTC m=+1204.817112334" lastFinishedPulling="2025-12-06 07:13:20.543639506 +0000 UTC m=+1213.998500550" observedRunningTime="2025-12-06 07:13:22.354013504 +0000 UTC m=+1215.808874548" watchObservedRunningTime="2025-12-06 07:13:22.358003521 +0000 UTC m=+1215.812864565" Dec 06 07:13:22 crc kubenswrapper[4945]: I1206 07:13:22.545919 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" podStartSLOduration=5.926403605 podStartE2EDuration="51.545894751s" podCreationTimestamp="2025-12-06 07:12:31 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.138416069 +0000 UTC m=+1168.593277103" lastFinishedPulling="2025-12-06 07:13:20.757907205 +0000 UTC m=+1214.212768249" observedRunningTime="2025-12-06 07:13:22.497698776 +0000 UTC m=+1215.952559840" watchObservedRunningTime="2025-12-06 07:13:22.545894751 +0000 UTC m=+1216.000756095" Dec 06 07:13:23 crc kubenswrapper[4945]: I1206 07:13:23.263936 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" event={"ID":"7305df73-ea7d-43f5-9119-9c12942796bc","Type":"ContainerStarted","Data":"d9e6c86650ba4c52fd7181ce4cd287c815189abfe004c641b9696e6436919d3f"} Dec 06 07:13:23 crc kubenswrapper[4945]: I1206 07:13:23.265371 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" Dec 06 07:13:23 crc kubenswrapper[4945]: I1206 07:13:23.268259 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" event={"ID":"6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc","Type":"ContainerStarted","Data":"eafcdf6347758f3ebae285a280fc7fb4dd5ed65d6425092cfe2293efb235304d"} Dec 06 07:13:23 crc kubenswrapper[4945]: I1206 07:13:23.268322 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" Dec 06 07:13:23 crc kubenswrapper[4945]: I1206 07:13:23.550101 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" podStartSLOduration=5.332209725 podStartE2EDuration="52.550078801s" podCreationTimestamp="2025-12-06 07:12:31 +0000 UTC" firstStartedPulling="2025-12-06 07:12:34.575315094 +0000 UTC m=+1168.030176138" lastFinishedPulling="2025-12-06 07:13:21.79318417 +0000 UTC m=+1215.248045214" observedRunningTime="2025-12-06 07:13:23.545386835 +0000 UTC m=+1217.000247899" watchObservedRunningTime="2025-12-06 07:13:23.550078801 +0000 UTC m=+1217.004939845" Dec 06 07:13:24 crc kubenswrapper[4945]: I1206 07:13:24.352616 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" event={"ID":"0e92db58-a1fc-4035-b635-f274d7411813","Type":"ContainerStarted","Data":"d7e033c02fd1acc1aea6bdb147cf64a32a10d3eb9d0fe610f579f7a7c3731e08"} Dec 06 07:13:24 crc kubenswrapper[4945]: I1206 07:13:24.354843 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" Dec 06 07:13:24 crc kubenswrapper[4945]: I1206 07:13:24.357511 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" event={"ID":"7c3f47b7-f1f2-408d-80d0-03a0ea9286ea","Type":"ContainerStarted","Data":"5746cbdfb090dca203287ab5854a31731ffc3f7a4227a21e81f656ec0f9e63fd"} Dec 06 07:13:24 crc kubenswrapper[4945]: I1206 07:13:24.357680 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" Dec 06 07:13:24 crc kubenswrapper[4945]: I1206 07:13:24.398000 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" podStartSLOduration=5.136152561 podStartE2EDuration="52.397981399s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:34.605081084 +0000 UTC m=+1168.059942128" lastFinishedPulling="2025-12-06 07:13:21.866909922 +0000 UTC m=+1215.321770966" observedRunningTime="2025-12-06 07:13:23.579332867 +0000 UTC m=+1217.034193911" watchObservedRunningTime="2025-12-06 07:13:24.397981399 +0000 UTC m=+1217.852842443" Dec 06 07:13:24 crc kubenswrapper[4945]: I1206 07:13:24.401544 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" podStartSLOduration=4.306053702 podStartE2EDuration="52.401526095s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.163108743 +0000 UTC m=+1168.617969787" lastFinishedPulling="2025-12-06 07:13:23.258581126 +0000 UTC m=+1216.713442180" observedRunningTime="2025-12-06 07:13:24.393425487 +0000 UTC m=+1217.848286531" watchObservedRunningTime="2025-12-06 07:13:24.401526095 +0000 UTC m=+1217.856387149" Dec 06 07:13:24 crc kubenswrapper[4945]: I1206 07:13:24.415805 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" podStartSLOduration=4.398045054 podStartE2EDuration="52.415783438s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.09825533 +0000 UTC m=+1168.553116374" lastFinishedPulling="2025-12-06 07:13:23.115993714 +0000 UTC m=+1216.570854758" observedRunningTime="2025-12-06 07:13:24.414148734 +0000 UTC m=+1217.869009778" watchObservedRunningTime="2025-12-06 07:13:24.415783438 +0000 UTC m=+1217.870644492" Dec 06 07:13:28 crc kubenswrapper[4945]: I1206 07:13:28.388049 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8" event={"ID":"4fe7c3eb-89a4-4935-8e19-935f98f8562b","Type":"ContainerStarted","Data":"d506b0ce725a80e2ff1c8a59f4abd45b71ba9693fcbd0e48468fc0a9886aa60e"} Dec 06 07:13:28 crc kubenswrapper[4945]: I1206 07:13:28.409465 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bd5w8" podStartSLOduration=2.61990811 podStartE2EDuration="55.409441745s" podCreationTimestamp="2025-12-06 07:12:33 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.218934573 +0000 UTC m=+1168.673795617" lastFinishedPulling="2025-12-06 07:13:28.008468208 +0000 UTC m=+1221.463329252" observedRunningTime="2025-12-06 07:13:28.401487151 +0000 UTC m=+1221.856348195" watchObservedRunningTime="2025-12-06 07:13:28.409441745 +0000 UTC m=+1221.864302799" Dec 06 07:13:28 crc kubenswrapper[4945]: I1206 07:13:28.934675 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56d49f" Dec 06 07:13:29 crc kubenswrapper[4945]: I1206 07:13:29.058178 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-f6b9t" Dec 06 07:13:32 crc kubenswrapper[4945]: I1206 07:13:32.367887 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kjxk4" Dec 06 07:13:32 crc kubenswrapper[4945]: I1206 07:13:32.520057 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wqq5g" Dec 06 07:13:32 crc kubenswrapper[4945]: I1206 07:13:32.543654 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bpsxr" Dec 06 07:13:32 crc kubenswrapper[4945]: I1206 07:13:32.654231 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cchx8" Dec 06 07:13:32 crc kubenswrapper[4945]: I1206 07:13:32.763912 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-x98dd" Dec 06 07:13:32 crc kubenswrapper[4945]: I1206 07:13:32.797674 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-r7zm6" Dec 06 07:13:32 crc kubenswrapper[4945]: I1206 07:13:32.823694 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5njm2" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.011086 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c4q9q" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.068372 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lwkxm" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.423185 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-v5krd" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.427238 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" event={"ID":"8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd","Type":"ContainerStarted","Data":"c150fa9f7ba2168a05dbd7cfa453df574b461d84d2022c6ccc63ba283a77b196"} Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.427551 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.429149 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" event={"ID":"716df590-6583-4b4b-bfdc-4d6bb2ac05a4","Type":"ContainerStarted","Data":"60a6547dfc8640a349b8b5ae29a601e8426d64b174910cc19344dea6afa6a37d"} Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.429384 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.431030 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" event={"ID":"039d67fe-f691-49a7-8e38-4a8dc2a2227e","Type":"ContainerStarted","Data":"a63df8e7319fe90a5f5c3214c7e218b79ab8a23240d601c8e9bf975934945db2"} Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.431258 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.432660 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" event={"ID":"c925c68e-4099-403e-b8ba-84d2d9bc689e","Type":"ContainerStarted","Data":"cb3bd639ce57529ce817189299843e2d2e68bb964679d452d19f21f63c4eb7b6"} Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.432840 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.494616 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" podStartSLOduration=5.239862924 podStartE2EDuration="1m2.494594329s" podCreationTimestamp="2025-12-06 07:12:31 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.253503033 +0000 UTC m=+1168.708364077" lastFinishedPulling="2025-12-06 07:13:32.508234438 +0000 UTC m=+1225.963095482" observedRunningTime="2025-12-06 07:13:33.493491249 +0000 UTC m=+1226.948352293" watchObservedRunningTime="2025-12-06 07:13:33.494594329 +0000 UTC m=+1226.949455373" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.528720 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" podStartSLOduration=4.096978913 podStartE2EDuration="1m1.528698355s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.237371469 +0000 UTC m=+1168.692232513" lastFinishedPulling="2025-12-06 07:13:32.669090911 +0000 UTC m=+1226.123951955" observedRunningTime="2025-12-06 07:13:33.519067136 +0000 UTC m=+1226.973928200" watchObservedRunningTime="2025-12-06 07:13:33.528698355 +0000 UTC m=+1226.983559399" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.554823 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" podStartSLOduration=4.350330833 podStartE2EDuration="1m1.554805137s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.237322238 +0000 UTC m=+1168.692183282" lastFinishedPulling="2025-12-06 07:13:32.441796542 +0000 UTC m=+1225.896657586" observedRunningTime="2025-12-06 07:13:33.538129159 +0000 UTC m=+1226.992990203" watchObservedRunningTime="2025-12-06 07:13:33.554805137 +0000 UTC m=+1227.009666181" Dec 06 07:13:33 crc kubenswrapper[4945]: I1206 07:13:33.973181 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" podStartSLOduration=5.5116613690000005 podStartE2EDuration="1m2.973155211s" podCreationTimestamp="2025-12-06 07:12:31 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.218983475 +0000 UTC m=+1168.673844519" lastFinishedPulling="2025-12-06 07:13:32.680477317 +0000 UTC m=+1226.135338361" observedRunningTime="2025-12-06 07:13:33.576333555 +0000 UTC m=+1227.031194619" watchObservedRunningTime="2025-12-06 07:13:33.973155211 +0000 UTC m=+1227.428016255" Dec 06 07:13:38 crc kubenswrapper[4945]: I1206 07:13:38.481291 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" event={"ID":"c0722d35-7779-4073-851f-0171050b6141","Type":"ContainerStarted","Data":"be6eca55909f6c664daa4f0555f63ea72f87c1f1aef28dafdbd635242e638615"} Dec 06 07:13:38 crc kubenswrapper[4945]: I1206 07:13:38.795522 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:13:38 crc kubenswrapper[4945]: I1206 07:13:38.795586 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:13:39 crc kubenswrapper[4945]: I1206 07:13:39.489948 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" event={"ID":"9bfd7a39-d83d-42de-bb30-0d1dc17bedca","Type":"ContainerStarted","Data":"40ed3fe581e9efc27e4c783ad6210a21db56aa2bc49320d46e80fbf760c226e0"} Dec 06 07:13:39 crc kubenswrapper[4945]: I1206 07:13:39.490358 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" Dec 06 07:13:39 crc kubenswrapper[4945]: I1206 07:13:39.490597 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" Dec 06 07:13:39 crc kubenswrapper[4945]: I1206 07:13:39.509909 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" podStartSLOduration=8.372075475 podStartE2EDuration="1m7.509892403s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.25377081 +0000 UTC m=+1168.708631854" lastFinishedPulling="2025-12-06 07:13:34.391587738 +0000 UTC m=+1227.846448782" observedRunningTime="2025-12-06 07:13:39.505900195 +0000 UTC m=+1232.960761259" watchObservedRunningTime="2025-12-06 07:13:39.509892403 +0000 UTC m=+1232.964753447" Dec 06 07:13:39 crc kubenswrapper[4945]: I1206 07:13:39.523693 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" podStartSLOduration=3.697306138 podStartE2EDuration="1m7.523674936s" podCreationTimestamp="2025-12-06 07:12:32 +0000 UTC" firstStartedPulling="2025-12-06 07:12:35.239452425 +0000 UTC m=+1168.694313469" lastFinishedPulling="2025-12-06 07:13:39.065821223 +0000 UTC m=+1232.520682267" observedRunningTime="2025-12-06 07:13:39.521377074 +0000 UTC m=+1232.976238148" watchObservedRunningTime="2025-12-06 07:13:39.523674936 +0000 UTC m=+1232.978535980" Dec 06 07:13:42 crc kubenswrapper[4945]: I1206 07:13:42.585221 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-cgbvh" Dec 06 07:13:42 crc kubenswrapper[4945]: I1206 07:13:42.681514 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-n8gdh" Dec 06 07:13:43 crc kubenswrapper[4945]: I1206 07:13:43.166075 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7cn4x" Dec 06 07:13:43 crc kubenswrapper[4945]: I1206 07:13:43.258716 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6tvc7" Dec 06 07:13:43 crc kubenswrapper[4945]: I1206 07:13:43.458623 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2vzqx" Dec 06 07:13:52 crc kubenswrapper[4945]: I1206 07:13:52.945450 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-pd42f" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.781128 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-4ccdj"] Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.783195 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.787104 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.787207 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.787333 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.797781 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-l6wc9" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.798499 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-4ccdj"] Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.812744 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-567c455747-47bhv"] Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.814399 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.824779 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.857458 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-47bhv"] Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.950937 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-dns-svc\") pod \"dnsmasq-dns-567c455747-47bhv\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.950990 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zstx\" (UniqueName: \"kubernetes.io/projected/7b898f33-7051-403c-96dc-2ae05176981b-kube-api-access-9zstx\") pod \"dnsmasq-dns-5cd484bb89-4ccdj\" (UID: \"7b898f33-7051-403c-96dc-2ae05176981b\") " pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.951024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb989\" (UniqueName: \"kubernetes.io/projected/d8f23c64-d522-4014-9cf1-7bff4e7230f2-kube-api-access-gb989\") pod \"dnsmasq-dns-567c455747-47bhv\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.951089 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b898f33-7051-403c-96dc-2ae05176981b-config\") pod \"dnsmasq-dns-5cd484bb89-4ccdj\" (UID: \"7b898f33-7051-403c-96dc-2ae05176981b\") " pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" Dec 06 07:14:06 crc kubenswrapper[4945]: I1206 07:14:06.951150 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-config\") pod \"dnsmasq-dns-567c455747-47bhv\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.052859 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b898f33-7051-403c-96dc-2ae05176981b-config\") pod \"dnsmasq-dns-5cd484bb89-4ccdj\" (UID: \"7b898f33-7051-403c-96dc-2ae05176981b\") " pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.053129 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-config\") pod \"dnsmasq-dns-567c455747-47bhv\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.053168 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-dns-svc\") pod \"dnsmasq-dns-567c455747-47bhv\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.053196 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zstx\" (UniqueName: \"kubernetes.io/projected/7b898f33-7051-403c-96dc-2ae05176981b-kube-api-access-9zstx\") pod \"dnsmasq-dns-5cd484bb89-4ccdj\" (UID: \"7b898f33-7051-403c-96dc-2ae05176981b\") " pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.053220 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb989\" (UniqueName: \"kubernetes.io/projected/d8f23c64-d522-4014-9cf1-7bff4e7230f2-kube-api-access-gb989\") pod \"dnsmasq-dns-567c455747-47bhv\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.055331 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.055376 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.064240 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b898f33-7051-403c-96dc-2ae05176981b-config\") pod \"dnsmasq-dns-5cd484bb89-4ccdj\" (UID: \"7b898f33-7051-403c-96dc-2ae05176981b\") " pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.064869 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-config\") pod \"dnsmasq-dns-567c455747-47bhv\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.066908 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-dns-svc\") pod \"dnsmasq-dns-567c455747-47bhv\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.067871 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.073387 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.082303 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb989\" (UniqueName: \"kubernetes.io/projected/d8f23c64-d522-4014-9cf1-7bff4e7230f2-kube-api-access-gb989\") pod \"dnsmasq-dns-567c455747-47bhv\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.085260 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zstx\" (UniqueName: \"kubernetes.io/projected/7b898f33-7051-403c-96dc-2ae05176981b-kube-api-access-9zstx\") pod \"dnsmasq-dns-5cd484bb89-4ccdj\" (UID: \"7b898f33-7051-403c-96dc-2ae05176981b\") " pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.126790 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-l6wc9" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.136030 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.147083 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.857842 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-4ccdj"] Dec 06 07:14:07 crc kubenswrapper[4945]: W1206 07:14:07.863787 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b898f33_7051_403c_96dc_2ae05176981b.slice/crio-019e555ca2fe24c0fdffdf076f1bf6dbc43057c448c422f8f22f5dcb400a028f WatchSource:0}: Error finding container 019e555ca2fe24c0fdffdf076f1bf6dbc43057c448c422f8f22f5dcb400a028f: Status 404 returned error can't find the container with id 019e555ca2fe24c0fdffdf076f1bf6dbc43057c448c422f8f22f5dcb400a028f Dec 06 07:14:07 crc kubenswrapper[4945]: I1206 07:14:07.885464 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-47bhv"] Dec 06 07:14:07 crc kubenswrapper[4945]: W1206 07:14:07.887628 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8f23c64_d522_4014_9cf1_7bff4e7230f2.slice/crio-a10878d10956feb51609a286fe4678191cd76490f66ed9816f227b7a0bac8c7a WatchSource:0}: Error finding container a10878d10956feb51609a286fe4678191cd76490f66ed9816f227b7a0bac8c7a: Status 404 returned error can't find the container with id a10878d10956feb51609a286fe4678191cd76490f66ed9816f227b7a0bac8c7a Dec 06 07:14:08 crc kubenswrapper[4945]: I1206 07:14:08.720133 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" event={"ID":"7b898f33-7051-403c-96dc-2ae05176981b","Type":"ContainerStarted","Data":"019e555ca2fe24c0fdffdf076f1bf6dbc43057c448c422f8f22f5dcb400a028f"} Dec 06 07:14:08 crc kubenswrapper[4945]: I1206 07:14:08.722656 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-47bhv" event={"ID":"d8f23c64-d522-4014-9cf1-7bff4e7230f2","Type":"ContainerStarted","Data":"a10878d10956feb51609a286fe4678191cd76490f66ed9816f227b7a0bac8c7a"} Dec 06 07:14:08 crc kubenswrapper[4945]: I1206 07:14:08.806213 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:14:08 crc kubenswrapper[4945]: I1206 07:14:08.806330 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:14:08 crc kubenswrapper[4945]: I1206 07:14:08.806407 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:14:08 crc kubenswrapper[4945]: I1206 07:14:08.807503 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd69474a2e644df9791d0d3744c09cd9214dcea77b7b07f5d390844d87230bc1"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:14:08 crc kubenswrapper[4945]: I1206 07:14:08.807589 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://fd69474a2e644df9791d0d3744c09cd9214dcea77b7b07f5d390844d87230bc1" gracePeriod=600 Dec 06 07:14:09 crc kubenswrapper[4945]: I1206 07:14:09.815991 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-47bhv"] Dec 06 07:14:09 crc kubenswrapper[4945]: I1206 07:14:09.838870 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="fd69474a2e644df9791d0d3744c09cd9214dcea77b7b07f5d390844d87230bc1" exitCode=0 Dec 06 07:14:09 crc kubenswrapper[4945]: I1206 07:14:09.839107 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"fd69474a2e644df9791d0d3744c09cd9214dcea77b7b07f5d390844d87230bc1"} Dec 06 07:14:09 crc kubenswrapper[4945]: I1206 07:14:09.839186 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"6db8402445675a35c1cfea5f14ca674ed295b8ef4463d76bc429b4cab4e4331a"} Dec 06 07:14:09 crc kubenswrapper[4945]: I1206 07:14:09.839315 4945 scope.go:117] "RemoveContainer" containerID="fcc05253d8d77edb32c43d3ca8c759b1e83d7e91658c6a423615205b104949ce" Dec 06 07:14:09 crc kubenswrapper[4945]: I1206 07:14:09.845501 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-tgskj"] Dec 06 07:14:09 crc kubenswrapper[4945]: I1206 07:14:09.847743 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:09 crc kubenswrapper[4945]: I1206 07:14:09.885029 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-tgskj"] Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.033084 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-dns-svc\") pod \"dnsmasq-dns-bc4b48fc9-tgskj\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.033718 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99qrf\" (UniqueName: \"kubernetes.io/projected/faa6de41-a4ad-4b26-8702-a90d9ff156bc-kube-api-access-99qrf\") pod \"dnsmasq-dns-bc4b48fc9-tgskj\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.033812 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-config\") pod \"dnsmasq-dns-bc4b48fc9-tgskj\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.135272 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-dns-svc\") pod \"dnsmasq-dns-bc4b48fc9-tgskj\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.135366 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99qrf\" (UniqueName: \"kubernetes.io/projected/faa6de41-a4ad-4b26-8702-a90d9ff156bc-kube-api-access-99qrf\") pod \"dnsmasq-dns-bc4b48fc9-tgskj\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.135405 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-config\") pod \"dnsmasq-dns-bc4b48fc9-tgskj\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.136485 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-config\") pod \"dnsmasq-dns-bc4b48fc9-tgskj\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.137094 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-dns-svc\") pod \"dnsmasq-dns-bc4b48fc9-tgskj\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.172982 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99qrf\" (UniqueName: \"kubernetes.io/projected/faa6de41-a4ad-4b26-8702-a90d9ff156bc-kube-api-access-99qrf\") pod \"dnsmasq-dns-bc4b48fc9-tgskj\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.180970 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-4ccdj"] Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.212679 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.216551 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb666b895-5bqzf"] Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.218176 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.251535 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-5bqzf"] Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.392263 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-dns-svc\") pod \"dnsmasq-dns-cb666b895-5bqzf\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.392357 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jfc5\" (UniqueName: \"kubernetes.io/projected/07a57f15-7d0c-468e-9e13-07bf8f493f63-kube-api-access-8jfc5\") pod \"dnsmasq-dns-cb666b895-5bqzf\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.392383 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-config\") pod \"dnsmasq-dns-cb666b895-5bqzf\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.494092 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-dns-svc\") pod \"dnsmasq-dns-cb666b895-5bqzf\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.494180 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jfc5\" (UniqueName: \"kubernetes.io/projected/07a57f15-7d0c-468e-9e13-07bf8f493f63-kube-api-access-8jfc5\") pod \"dnsmasq-dns-cb666b895-5bqzf\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.494215 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-config\") pod \"dnsmasq-dns-cb666b895-5bqzf\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.495157 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-dns-svc\") pod \"dnsmasq-dns-cb666b895-5bqzf\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.496245 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-config\") pod \"dnsmasq-dns-cb666b895-5bqzf\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.530262 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jfc5\" (UniqueName: \"kubernetes.io/projected/07a57f15-7d0c-468e-9e13-07bf8f493f63-kube-api-access-8jfc5\") pod \"dnsmasq-dns-cb666b895-5bqzf\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.598025 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:10 crc kubenswrapper[4945]: I1206 07:14:10.864881 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-tgskj"] Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.045837 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.048512 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.057102 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-g24n6" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.057309 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.057415 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.057524 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.057634 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.057797 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.058835 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.103263 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.151940 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-5bqzf"] Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213538 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213622 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213654 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213703 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d71d0084-b4bc-4be5-971b-eced6e6d66ea-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213730 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d5hl\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-kube-api-access-6d5hl\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213772 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d71d0084-b4bc-4be5-971b-eced6e6d66ea-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213795 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213859 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213881 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213909 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.213941 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.314765 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.314830 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.314858 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.314885 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.314908 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.314923 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.314941 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.314970 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d71d0084-b4bc-4be5-971b-eced6e6d66ea-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.314989 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d5hl\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-kube-api-access-6d5hl\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.315015 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d71d0084-b4bc-4be5-971b-eced6e6d66ea-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.315031 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.316298 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.316568 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.317237 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.317252 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.317496 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.319326 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.324198 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.329114 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.331087 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d71d0084-b4bc-4be5-971b-eced6e6d66ea-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.331110 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d71d0084-b4bc-4be5-971b-eced6e6d66ea-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.336707 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d5hl\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-kube-api-access-6d5hl\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.343303 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.344608 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.352991 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.353136 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.353228 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.353419 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.353513 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-tdw59" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.353603 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.353718 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.360685 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.361981 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.414101 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.517797 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.517900 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.517924 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/250c85e2-e14d-4f93-822c-c9a88352eafe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.517965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.517984 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.518023 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.518074 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/250c85e2-e14d-4f93-822c-c9a88352eafe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.518096 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.518114 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p69tr\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-kube-api-access-p69tr\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.519110 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.519539 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.622242 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.622637 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p69tr\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-kube-api-access-p69tr\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.622689 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.622706 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.622729 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.622773 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.622788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/250c85e2-e14d-4f93-822c-c9a88352eafe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.623107 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.623169 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.623259 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.623345 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/250c85e2-e14d-4f93-822c-c9a88352eafe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.623714 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.624044 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.624929 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.629344 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/250c85e2-e14d-4f93-822c-c9a88352eafe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.629566 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.629649 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.630411 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.647508 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.650477 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/250c85e2-e14d-4f93-822c-c9a88352eafe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.650670 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.651949 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.661326 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p69tr\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-kube-api-access-p69tr\") pod \"rabbitmq-cell1-server-0\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.727076 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.886899 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" event={"ID":"faa6de41-a4ad-4b26-8702-a90d9ff156bc","Type":"ContainerStarted","Data":"38890cf88e1d6fb147c4b6f11d50ed3d995352f3651048b29847243eb9fe44a9"} Dec 06 07:14:11 crc kubenswrapper[4945]: I1206 07:14:11.888928 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-5bqzf" event={"ID":"07a57f15-7d0c-468e-9e13-07bf8f493f63","Type":"ContainerStarted","Data":"e949e2747a3c292075ab00031002ad15ac7719a2b732efea2a9c31e1593c73d9"} Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.089264 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 07:14:12 crc kubenswrapper[4945]: W1206 07:14:12.104629 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd71d0084_b4bc_4be5_971b_eced6e6d66ea.slice/crio-44bc0c48bbbaf35075c884c6fb25ca15b8147b4de1fb584d714e11d7ffaf858b WatchSource:0}: Error finding container 44bc0c48bbbaf35075c884c6fb25ca15b8147b4de1fb584d714e11d7ffaf858b: Status 404 returned error can't find the container with id 44bc0c48bbbaf35075c884c6fb25ca15b8147b4de1fb584d714e11d7ffaf858b Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.247964 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.249961 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.253760 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-stf7c" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.253987 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.256365 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.256569 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.262437 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.266989 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.367970 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.368019 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.368072 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.368107 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.368142 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-default\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.368161 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2twnc\" (UniqueName: \"kubernetes.io/projected/fe30cfa7-9746-4771-bac5-29e3a3662aec-kube-api-access-2twnc\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.368191 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.368214 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-kolla-config\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.473011 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-default\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.473054 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2twnc\" (UniqueName: \"kubernetes.io/projected/fe30cfa7-9746-4771-bac5-29e3a3662aec-kube-api-access-2twnc\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.473084 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.473113 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-kolla-config\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.473141 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.473162 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.473201 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.473245 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.474925 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-kolla-config\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.476200 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-default\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.476821 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.478811 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.481814 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.507423 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.512079 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.512399 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.517406 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2twnc\" (UniqueName: \"kubernetes.io/projected/fe30cfa7-9746-4771-bac5-29e3a3662aec-kube-api-access-2twnc\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.517508 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.580336 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.925950 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d71d0084-b4bc-4be5-971b-eced6e6d66ea","Type":"ContainerStarted","Data":"44bc0c48bbbaf35075c884c6fb25ca15b8147b4de1fb584d714e11d7ffaf858b"} Dec 06 07:14:12 crc kubenswrapper[4945]: I1206 07:14:12.927428 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"250c85e2-e14d-4f93-822c-c9a88352eafe","Type":"ContainerStarted","Data":"97d9f52a130ec28dbe4efd27747eb9d28f6506c702f7521e8b4f6a894d1b9c21"} Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.227511 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.471104 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.472549 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.475242 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.476757 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.476935 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-g4tjp" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.477517 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.486834 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.607904 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.607969 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.608002 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44pbb\" (UniqueName: \"kubernetes.io/projected/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kube-api-access-44pbb\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.608053 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.608087 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.608132 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.608162 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.608189 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.715035 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.715085 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.715115 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44pbb\" (UniqueName: \"kubernetes.io/projected/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kube-api-access-44pbb\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.715155 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.715182 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.715224 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.715245 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.715266 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.716949 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.717412 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.727974 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.728586 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.728623 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.729151 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.730547 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.774515 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.775152 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44pbb\" (UniqueName: \"kubernetes.io/projected/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kube-api-access-44pbb\") pod \"openstack-cell1-galera-0\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.834598 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 07:14:13 crc kubenswrapper[4945]: I1206 07:14:13.950893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fe30cfa7-9746-4771-bac5-29e3a3662aec","Type":"ContainerStarted","Data":"2a0b523809cf4ba71213abf525a2e8725ce5614df35fd12c49111e1d3470477c"} Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.391620 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.393072 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.397738 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.398046 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.398190 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-lmlwb" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.446437 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.483604 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q8js\" (UniqueName: \"kubernetes.io/projected/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kube-api-access-8q8js\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.484982 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kolla-config\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.485766 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.485832 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.485991 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-config-data\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.592776 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.592867 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.592893 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-config-data\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.593077 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q8js\" (UniqueName: \"kubernetes.io/projected/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kube-api-access-8q8js\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.593205 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kolla-config\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.594323 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kolla-config\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.594677 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-config-data\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.620066 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.667077 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q8js\" (UniqueName: \"kubernetes.io/projected/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kube-api-access-8q8js\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.689845 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " pod="openstack/memcached-0" Dec 06 07:14:15 crc kubenswrapper[4945]: I1206 07:14:15.812755 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 07:14:17 crc kubenswrapper[4945]: I1206 07:14:17.264910 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:14:17 crc kubenswrapper[4945]: I1206 07:14:17.268695 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 07:14:17 crc kubenswrapper[4945]: I1206 07:14:17.275050 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-9v5f4" Dec 06 07:14:17 crc kubenswrapper[4945]: I1206 07:14:17.278998 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:14:17 crc kubenswrapper[4945]: I1206 07:14:17.374082 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9w2f\" (UniqueName: \"kubernetes.io/projected/f5f1ef51-68e5-4133-a248-1d7251b5bb96-kube-api-access-z9w2f\") pod \"kube-state-metrics-0\" (UID: \"f5f1ef51-68e5-4133-a248-1d7251b5bb96\") " pod="openstack/kube-state-metrics-0" Dec 06 07:14:17 crc kubenswrapper[4945]: I1206 07:14:17.475751 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9w2f\" (UniqueName: \"kubernetes.io/projected/f5f1ef51-68e5-4133-a248-1d7251b5bb96-kube-api-access-z9w2f\") pod \"kube-state-metrics-0\" (UID: \"f5f1ef51-68e5-4133-a248-1d7251b5bb96\") " pod="openstack/kube-state-metrics-0" Dec 06 07:14:17 crc kubenswrapper[4945]: I1206 07:14:17.525571 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9w2f\" (UniqueName: \"kubernetes.io/projected/f5f1ef51-68e5-4133-a248-1d7251b5bb96-kube-api-access-z9w2f\") pod \"kube-state-metrics-0\" (UID: \"f5f1ef51-68e5-4133-a248-1d7251b5bb96\") " pod="openstack/kube-state-metrics-0" Dec 06 07:14:17 crc kubenswrapper[4945]: I1206 07:14:17.609728 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 07:14:18 crc kubenswrapper[4945]: I1206 07:14:18.991094 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 07:14:18 crc kubenswrapper[4945]: I1206 07:14:18.992466 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:18.996668 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:18.997027 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:18.997198 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:18.997332 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-dmkf2" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:18.999783 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.015554 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.069887 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.069949 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.070024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.070119 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.070159 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqbgl\" (UniqueName: \"kubernetes.io/projected/d5496b71-8389-4d16-97c0-5752a2f1ef2a-kube-api-access-xqbgl\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.070199 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-config\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.070260 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.070474 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.172966 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.173036 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.173097 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.173161 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.173197 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqbgl\" (UniqueName: \"kubernetes.io/projected/d5496b71-8389-4d16-97c0-5752a2f1ef2a-kube-api-access-xqbgl\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.173239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-config\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.173269 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.174415 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.174761 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.176079 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.176656 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.177301 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-config\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.184369 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.200058 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.203651 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqbgl\" (UniqueName: \"kubernetes.io/projected/d5496b71-8389-4d16-97c0-5752a2f1ef2a-kube-api-access-xqbgl\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.215130 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.218118 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:19 crc kubenswrapper[4945]: I1206 07:14:19.325271 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.787341 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xc2dd"] Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.788797 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.792865 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-lknnz" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.792970 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.793189 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.800538 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-2996k"] Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.802717 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.809781 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xc2dd"] Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.829601 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2996k"] Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.912850 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-run\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.912899 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run-ovn\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.912928 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-scripts\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.912989 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-log-ovn\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.913016 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-combined-ca-bundle\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.913137 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bl4s\" (UniqueName: \"kubernetes.io/projected/9c092be9-37db-49bb-be6b-56ce2e0e15fd-kube-api-access-5bl4s\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.913239 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-etc-ovs\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.913294 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.913345 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-lib\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.913368 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-log\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.913406 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkrf2\" (UniqueName: \"kubernetes.io/projected/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-kube-api-access-pkrf2\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.913456 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-ovn-controller-tls-certs\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:20 crc kubenswrapper[4945]: I1206 07:14:20.913520 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c092be9-37db-49bb-be6b-56ce2e0e15fd-scripts\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.014707 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.014763 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-lib\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.014788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-log\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.014817 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkrf2\" (UniqueName: \"kubernetes.io/projected/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-kube-api-access-pkrf2\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.014850 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-ovn-controller-tls-certs\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.014888 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c092be9-37db-49bb-be6b-56ce2e0e15fd-scripts\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.014911 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-run\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.014935 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run-ovn\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015003 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-scripts\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015081 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-log-ovn\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015111 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-combined-ca-bundle\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015139 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bl4s\" (UniqueName: \"kubernetes.io/projected/9c092be9-37db-49bb-be6b-56ce2e0e15fd-kube-api-access-5bl4s\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015183 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-etc-ovs\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015458 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015513 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-lib\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015589 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-run\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015622 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-etc-ovs\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015694 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-log\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015871 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run-ovn\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.015985 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-log-ovn\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.018372 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-scripts\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.018809 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c092be9-37db-49bb-be6b-56ce2e0e15fd-scripts\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.021496 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-ovn-controller-tls-certs\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.038617 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bl4s\" (UniqueName: \"kubernetes.io/projected/9c092be9-37db-49bb-be6b-56ce2e0e15fd-kube-api-access-5bl4s\") pod \"ovn-controller-ovs-2996k\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.045343 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkrf2\" (UniqueName: \"kubernetes.io/projected/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-kube-api-access-pkrf2\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.057793 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-combined-ca-bundle\") pod \"ovn-controller-xc2dd\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.107266 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xc2dd" Dec 06 07:14:21 crc kubenswrapper[4945]: I1206 07:14:21.131779 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:14:22 crc kubenswrapper[4945]: I1206 07:14:22.918374 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 07:14:22 crc kubenswrapper[4945]: I1206 07:14:22.920834 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:22 crc kubenswrapper[4945]: I1206 07:14:22.962951 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 06 07:14:22 crc kubenswrapper[4945]: I1206 07:14:22.965639 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-f6b79" Dec 06 07:14:22 crc kubenswrapper[4945]: I1206 07:14:22.965892 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 06 07:14:22 crc kubenswrapper[4945]: I1206 07:14:22.966046 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.005489 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.098487 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbjr4\" (UniqueName: \"kubernetes.io/projected/b51cdd1a-131a-4f4c-8521-93890fd48722-kube-api-access-nbjr4\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.098573 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.098639 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.098662 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-config\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.098683 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.098759 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.098811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.098857 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.199854 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.199911 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbjr4\" (UniqueName: \"kubernetes.io/projected/b51cdd1a-131a-4f4c-8521-93890fd48722-kube-api-access-nbjr4\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.199939 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.199991 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.200014 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-config\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.200032 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.200081 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.200113 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.200351 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.200628 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.201347 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-config\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.201522 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.204576 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.205079 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.209964 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.233628 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.263770 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbjr4\" (UniqueName: \"kubernetes.io/projected/b51cdd1a-131a-4f4c-8521-93890fd48722-kube-api-access-nbjr4\") pod \"ovsdbserver-sb-0\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:23 crc kubenswrapper[4945]: I1206 07:14:23.274520 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 07:14:34 crc kubenswrapper[4945]: E1206 07:14:34.202003 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d" Dec 06 07:14:34 crc kubenswrapper[4945]: E1206 07:14:34.202927 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p69tr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(250c85e2-e14d-4f93-822c-c9a88352eafe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:14:34 crc kubenswrapper[4945]: E1206 07:14:34.204785 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="250c85e2-e14d-4f93-822c-c9a88352eafe" Dec 06 07:14:34 crc kubenswrapper[4945]: E1206 07:14:34.242781 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d" Dec 06 07:14:34 crc kubenswrapper[4945]: E1206 07:14:34.243047 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6d5hl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(d71d0084-b4bc-4be5-971b-eced6e6d66ea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:14:34 crc kubenswrapper[4945]: E1206 07:14:34.246123 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" Dec 06 07:14:34 crc kubenswrapper[4945]: E1206 07:14:34.279813 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="250c85e2-e14d-4f93-822c-c9a88352eafe" Dec 06 07:14:34 crc kubenswrapper[4945]: E1206 07:14:34.281095 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d\\\"\"" pod="openstack/rabbitmq-server-0" podUID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" Dec 06 07:14:49 crc kubenswrapper[4945]: E1206 07:14:49.264591 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a" Dec 06 07:14:49 crc kubenswrapper[4945]: E1206 07:14:49.265742 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2twnc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(fe30cfa7-9746-4771-bac5-29e3a3662aec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:14:49 crc kubenswrapper[4945]: E1206 07:14:49.269578 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="fe30cfa7-9746-4771-bac5-29e3a3662aec" Dec 06 07:14:49 crc kubenswrapper[4945]: E1206 07:14:49.472581 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a\\\"\"" pod="openstack/openstack-galera-0" podUID="fe30cfa7-9746-4771-bac5-29e3a3662aec" Dec 06 07:14:49 crc kubenswrapper[4945]: I1206 07:14:49.828423 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2996k"] Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.677698 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.678121 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-99qrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-bc4b48fc9-tgskj_openstack(faa6de41-a4ad-4b26-8702-a90d9ff156bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.679332 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" podUID="faa6de41-a4ad-4b26-8702-a90d9ff156bc" Dec 06 07:14:50 crc kubenswrapper[4945]: W1206 07:14:50.689110 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c092be9_37db_49bb_be6b_56ce2e0e15fd.slice/crio-f5715ee34ed4cb57a22cbe9aa3687f0bce4b9d859a4437c104d0b5751af78ab7 WatchSource:0}: Error finding container f5715ee34ed4cb57a22cbe9aa3687f0bce4b9d859a4437c104d0b5751af78ab7: Status 404 returned error can't find the container with id f5715ee34ed4cb57a22cbe9aa3687f0bce4b9d859a4437c104d0b5751af78ab7 Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.703023 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.703202 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8jfc5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-cb666b895-5bqzf_openstack(07a57f15-7d0c-468e-9e13-07bf8f493f63): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.705030 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-cb666b895-5bqzf" podUID="07a57f15-7d0c-468e-9e13-07bf8f493f63" Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.733440 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.733636 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gb989,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-567c455747-47bhv_openstack(d8f23c64-d522-4014-9cf1-7bff4e7230f2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.738396 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-567c455747-47bhv" podUID="d8f23c64-d522-4014-9cf1-7bff4e7230f2" Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.738620 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.738788 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9zstx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5cd484bb89-4ccdj_openstack(7b898f33-7051-403c-96dc-2ae05176981b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:14:50 crc kubenswrapper[4945]: E1206 07:14:50.740914 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" podUID="7b898f33-7051-403c-96dc-2ae05176981b" Dec 06 07:14:51 crc kubenswrapper[4945]: I1206 07:14:51.318433 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xc2dd"] Dec 06 07:14:51 crc kubenswrapper[4945]: W1206 07:14:51.326884 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fb32cca_c8f2_4ff6_a05c_25f3e56b5478.slice/crio-e49cafcdd2705b328ea643d4240d8c3a8df1612a9c54575dfeb9cfa088329fc5 WatchSource:0}: Error finding container e49cafcdd2705b328ea643d4240d8c3a8df1612a9c54575dfeb9cfa088329fc5: Status 404 returned error can't find the container with id e49cafcdd2705b328ea643d4240d8c3a8df1612a9c54575dfeb9cfa088329fc5 Dec 06 07:14:51 crc kubenswrapper[4945]: W1206 07:14:51.334503 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae8dae45_f01e_4ef0_8814_07511ff2fcad.slice/crio-4a4f53aad4d8b8ad07a84460aeafda04a68f1d5b681c2c2d550714a3c60a4766 WatchSource:0}: Error finding container 4a4f53aad4d8b8ad07a84460aeafda04a68f1d5b681c2c2d550714a3c60a4766: Status 404 returned error can't find the container with id 4a4f53aad4d8b8ad07a84460aeafda04a68f1d5b681c2c2d550714a3c60a4766 Dec 06 07:14:51 crc kubenswrapper[4945]: I1206 07:14:51.339604 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 07:14:51 crc kubenswrapper[4945]: I1206 07:14:51.348385 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:14:51 crc kubenswrapper[4945]: I1206 07:14:51.432968 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 07:14:51 crc kubenswrapper[4945]: W1206 07:14:51.438646 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef14eff5_a3d8_48a8_b3fc_b59b97f68b67.slice/crio-66c502aefd40a1a3a51bb03f1f9bedb664bf396043e3ee36b232297da1b7cf8f WatchSource:0}: Error finding container 66c502aefd40a1a3a51bb03f1f9bedb664bf396043e3ee36b232297da1b7cf8f: Status 404 returned error can't find the container with id 66c502aefd40a1a3a51bb03f1f9bedb664bf396043e3ee36b232297da1b7cf8f Dec 06 07:14:51 crc kubenswrapper[4945]: I1206 07:14:51.502653 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ae8dae45-f01e-4ef0-8814-07511ff2fcad","Type":"ContainerStarted","Data":"4a4f53aad4d8b8ad07a84460aeafda04a68f1d5b681c2c2d550714a3c60a4766"} Dec 06 07:14:51 crc kubenswrapper[4945]: I1206 07:14:51.504578 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f5f1ef51-68e5-4133-a248-1d7251b5bb96","Type":"ContainerStarted","Data":"8d0c5e2b19aa7a04906ff8be06ade8102a376be71b052d6e55158fe1764151ea"} Dec 06 07:14:51 crc kubenswrapper[4945]: I1206 07:14:51.509924 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xc2dd" event={"ID":"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478","Type":"ContainerStarted","Data":"e49cafcdd2705b328ea643d4240d8c3a8df1612a9c54575dfeb9cfa088329fc5"} Dec 06 07:14:51 crc kubenswrapper[4945]: I1206 07:14:51.517091 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2996k" event={"ID":"9c092be9-37db-49bb-be6b-56ce2e0e15fd","Type":"ContainerStarted","Data":"f5715ee34ed4cb57a22cbe9aa3687f0bce4b9d859a4437c104d0b5751af78ab7"} Dec 06 07:14:51 crc kubenswrapper[4945]: I1206 07:14:51.517835 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 07:14:51 crc kubenswrapper[4945]: I1206 07:14:51.518716 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67","Type":"ContainerStarted","Data":"66c502aefd40a1a3a51bb03f1f9bedb664bf396043e3ee36b232297da1b7cf8f"} Dec 06 07:14:51 crc kubenswrapper[4945]: E1206 07:14:51.522514 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792\\\"\"" pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" podUID="faa6de41-a4ad-4b26-8702-a90d9ff156bc" Dec 06 07:14:51 crc kubenswrapper[4945]: E1206 07:14:51.522514 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792\\\"\"" pod="openstack/dnsmasq-dns-cb666b895-5bqzf" podUID="07a57f15-7d0c-468e-9e13-07bf8f493f63" Dec 06 07:14:51 crc kubenswrapper[4945]: W1206 07:14:51.525517 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5496b71_8389_4d16_97c0_5752a2f1ef2a.slice/crio-48a0d29b488df48720b7b28465ae6845e51f0c383c1cee85f2166c39938f50df WatchSource:0}: Error finding container 48a0d29b488df48720b7b28465ae6845e51f0c383c1cee85f2166c39938f50df: Status 404 returned error can't find the container with id 48a0d29b488df48720b7b28465ae6845e51f0c383c1cee85f2166c39938f50df Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.185145 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.282414 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.292301 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" Dec 06 07:14:52 crc kubenswrapper[4945]: W1206 07:14:52.337960 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb51cdd1a_131a_4f4c_8521_93890fd48722.slice/crio-822d32044760edff7f3133ad6da795c622a90d0cd824f06234ebbfefb18b49e0 WatchSource:0}: Error finding container 822d32044760edff7f3133ad6da795c622a90d0cd824f06234ebbfefb18b49e0: Status 404 returned error can't find the container with id 822d32044760edff7f3133ad6da795c622a90d0cd824f06234ebbfefb18b49e0 Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.379061 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b898f33-7051-403c-96dc-2ae05176981b-config\") pod \"7b898f33-7051-403c-96dc-2ae05176981b\" (UID: \"7b898f33-7051-403c-96dc-2ae05176981b\") " Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.379397 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-dns-svc\") pod \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.379447 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb989\" (UniqueName: \"kubernetes.io/projected/d8f23c64-d522-4014-9cf1-7bff4e7230f2-kube-api-access-gb989\") pod \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.379478 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zstx\" (UniqueName: \"kubernetes.io/projected/7b898f33-7051-403c-96dc-2ae05176981b-kube-api-access-9zstx\") pod \"7b898f33-7051-403c-96dc-2ae05176981b\" (UID: \"7b898f33-7051-403c-96dc-2ae05176981b\") " Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.379515 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-config\") pod \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\" (UID: \"d8f23c64-d522-4014-9cf1-7bff4e7230f2\") " Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.379937 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b898f33-7051-403c-96dc-2ae05176981b-config" (OuterVolumeSpecName: "config") pod "7b898f33-7051-403c-96dc-2ae05176981b" (UID: "7b898f33-7051-403c-96dc-2ae05176981b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.381076 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d8f23c64-d522-4014-9cf1-7bff4e7230f2" (UID: "d8f23c64-d522-4014-9cf1-7bff4e7230f2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.381172 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-config" (OuterVolumeSpecName: "config") pod "d8f23c64-d522-4014-9cf1-7bff4e7230f2" (UID: "d8f23c64-d522-4014-9cf1-7bff4e7230f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.386105 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b898f33-7051-403c-96dc-2ae05176981b-kube-api-access-9zstx" (OuterVolumeSpecName: "kube-api-access-9zstx") pod "7b898f33-7051-403c-96dc-2ae05176981b" (UID: "7b898f33-7051-403c-96dc-2ae05176981b"). InnerVolumeSpecName "kube-api-access-9zstx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.386982 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f23c64-d522-4014-9cf1-7bff4e7230f2-kube-api-access-gb989" (OuterVolumeSpecName: "kube-api-access-gb989") pod "d8f23c64-d522-4014-9cf1-7bff4e7230f2" (UID: "d8f23c64-d522-4014-9cf1-7bff4e7230f2"). InnerVolumeSpecName "kube-api-access-gb989". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.482123 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zstx\" (UniqueName: \"kubernetes.io/projected/7b898f33-7051-403c-96dc-2ae05176981b-kube-api-access-9zstx\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.482173 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.482186 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b898f33-7051-403c-96dc-2ae05176981b-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.482198 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8f23c64-d522-4014-9cf1-7bff4e7230f2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.482211 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb989\" (UniqueName: \"kubernetes.io/projected/d8f23c64-d522-4014-9cf1-7bff4e7230f2-kube-api-access-gb989\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.541383 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"250c85e2-e14d-4f93-822c-c9a88352eafe","Type":"ContainerStarted","Data":"3da187010c6d527c479f778bf05d2b504b2c0f1fcf3e34a35517eca9354e58fd"} Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.542594 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-47bhv" event={"ID":"d8f23c64-d522-4014-9cf1-7bff4e7230f2","Type":"ContainerDied","Data":"a10878d10956feb51609a286fe4678191cd76490f66ed9816f227b7a0bac8c7a"} Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.542669 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-47bhv" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.559616 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" event={"ID":"7b898f33-7051-403c-96dc-2ae05176981b","Type":"ContainerDied","Data":"019e555ca2fe24c0fdffdf076f1bf6dbc43057c448c422f8f22f5dcb400a028f"} Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.559638 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-4ccdj" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.562082 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d71d0084-b4bc-4be5-971b-eced6e6d66ea","Type":"ContainerStarted","Data":"299978e81ebf6ee60eff482d48ee097c771fe6fe34e8dc8d956786247712867f"} Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.569994 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d5496b71-8389-4d16-97c0-5752a2f1ef2a","Type":"ContainerStarted","Data":"48a0d29b488df48720b7b28465ae6845e51f0c383c1cee85f2166c39938f50df"} Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.572870 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b51cdd1a-131a-4f4c-8521-93890fd48722","Type":"ContainerStarted","Data":"822d32044760edff7f3133ad6da795c622a90d0cd824f06234ebbfefb18b49e0"} Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.666796 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-47bhv"] Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.675212 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-567c455747-47bhv"] Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.698580 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-4ccdj"] Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.704975 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-4ccdj"] Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.967554 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b898f33-7051-403c-96dc-2ae05176981b" path="/var/lib/kubelet/pods/7b898f33-7051-403c-96dc-2ae05176981b/volumes" Dec 06 07:14:52 crc kubenswrapper[4945]: I1206 07:14:52.968047 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8f23c64-d522-4014-9cf1-7bff4e7230f2" path="/var/lib/kubelet/pods/d8f23c64-d522-4014-9cf1-7bff4e7230f2/volumes" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.137946 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-lthjk"] Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.140863 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.143362 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.158900 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lthjk"] Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.212307 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-combined-ca-bundle\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.212372 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zllcm\" (UniqueName: \"kubernetes.io/projected/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-kube-api-access-zllcm\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.212413 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.212521 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovn-rundir\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.212584 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-config\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.212613 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovs-rundir\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.314573 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.314678 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovn-rundir\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.314724 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-config\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.314750 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovs-rundir\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.314775 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-combined-ca-bundle\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.314818 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zllcm\" (UniqueName: \"kubernetes.io/projected/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-kube-api-access-zllcm\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.315012 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovn-rundir\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.315094 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovs-rundir\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.315726 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-config\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.317258 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-tgskj"] Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.319648 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-combined-ca-bundle\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.319657 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.344607 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-bbc9w"] Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.346263 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zllcm\" (UniqueName: \"kubernetes.io/projected/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-kube-api-access-zllcm\") pod \"ovn-controller-metrics-lthjk\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.346646 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.350547 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.383911 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-bbc9w"] Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.473661 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.517559 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-config\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.517613 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k948d\" (UniqueName: \"kubernetes.io/projected/d175d796-a22c-43e0-af90-739e9dcfb044-kube-api-access-k948d\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.517721 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-ovsdbserver-nb\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.517763 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-dns-svc\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.625223 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k948d\" (UniqueName: \"kubernetes.io/projected/d175d796-a22c-43e0-af90-739e9dcfb044-kube-api-access-k948d\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.625700 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-ovsdbserver-nb\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.625747 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-dns-svc\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.625888 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-config\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.626846 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-ovsdbserver-nb\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.627358 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-dns-svc\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.627692 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-config\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.668637 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ae8dae45-f01e-4ef0-8814-07511ff2fcad","Type":"ContainerStarted","Data":"c7eb1e87ecd225ee2282907be972c030861b9852e609a8c7badc9b0c36c11a22"} Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.696906 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k948d\" (UniqueName: \"kubernetes.io/projected/d175d796-a22c-43e0-af90-739e9dcfb044-kube-api-access-k948d\") pod \"dnsmasq-dns-57db9b5bc9-bbc9w\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.727391 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-5bqzf"] Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.732650 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.763388 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-j4fkv"] Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.783712 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.790459 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.805371 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-j4fkv"] Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.932298 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnkbg\" (UniqueName: \"kubernetes.io/projected/961d0927-7c9d-4b8d-8e5d-a97883f66943-kube-api-access-pnkbg\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.932357 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.932405 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-config\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.932428 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:53 crc kubenswrapper[4945]: I1206 07:14:53.932488 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-dns-svc\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.034696 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnkbg\" (UniqueName: \"kubernetes.io/projected/961d0927-7c9d-4b8d-8e5d-a97883f66943-kube-api-access-pnkbg\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.034767 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.034833 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-config\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.034859 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.034996 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-dns-svc\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.038208 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.039335 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-config\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.039760 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-dns-svc\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.043115 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.066033 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnkbg\" (UniqueName: \"kubernetes.io/projected/961d0927-7c9d-4b8d-8e5d-a97883f66943-kube-api-access-pnkbg\") pod \"dnsmasq-dns-db7757ddc-j4fkv\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:54 crc kubenswrapper[4945]: I1206 07:14:54.141537 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.656140 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.663574 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.686181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" event={"ID":"faa6de41-a4ad-4b26-8702-a90d9ff156bc","Type":"ContainerDied","Data":"38890cf88e1d6fb147c4b6f11d50ed3d995352f3651048b29847243eb9fe44a9"} Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.686189 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-tgskj" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.687316 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-5bqzf" event={"ID":"07a57f15-7d0c-468e-9e13-07bf8f493f63","Type":"ContainerDied","Data":"e949e2747a3c292075ab00031002ad15ac7719a2b732efea2a9c31e1593c73d9"} Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.687360 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-5bqzf" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.762070 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99qrf\" (UniqueName: \"kubernetes.io/projected/faa6de41-a4ad-4b26-8702-a90d9ff156bc-kube-api-access-99qrf\") pod \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.762228 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-dns-svc\") pod \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.762370 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-config\") pod \"07a57f15-7d0c-468e-9e13-07bf8f493f63\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.762409 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-config\") pod \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\" (UID: \"faa6de41-a4ad-4b26-8702-a90d9ff156bc\") " Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.762509 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jfc5\" (UniqueName: \"kubernetes.io/projected/07a57f15-7d0c-468e-9e13-07bf8f493f63-kube-api-access-8jfc5\") pod \"07a57f15-7d0c-468e-9e13-07bf8f493f63\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.762629 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-dns-svc\") pod \"07a57f15-7d0c-468e-9e13-07bf8f493f63\" (UID: \"07a57f15-7d0c-468e-9e13-07bf8f493f63\") " Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.762984 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "faa6de41-a4ad-4b26-8702-a90d9ff156bc" (UID: "faa6de41-a4ad-4b26-8702-a90d9ff156bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.763075 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-config" (OuterVolumeSpecName: "config") pod "07a57f15-7d0c-468e-9e13-07bf8f493f63" (UID: "07a57f15-7d0c-468e-9e13-07bf8f493f63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.763157 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-config" (OuterVolumeSpecName: "config") pod "faa6de41-a4ad-4b26-8702-a90d9ff156bc" (UID: "faa6de41-a4ad-4b26-8702-a90d9ff156bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.763180 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.763165 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "07a57f15-7d0c-468e-9e13-07bf8f493f63" (UID: "07a57f15-7d0c-468e-9e13-07bf8f493f63"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.769556 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07a57f15-7d0c-468e-9e13-07bf8f493f63-kube-api-access-8jfc5" (OuterVolumeSpecName: "kube-api-access-8jfc5") pod "07a57f15-7d0c-468e-9e13-07bf8f493f63" (UID: "07a57f15-7d0c-468e-9e13-07bf8f493f63"). InnerVolumeSpecName "kube-api-access-8jfc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.769647 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faa6de41-a4ad-4b26-8702-a90d9ff156bc-kube-api-access-99qrf" (OuterVolumeSpecName: "kube-api-access-99qrf") pod "faa6de41-a4ad-4b26-8702-a90d9ff156bc" (UID: "faa6de41-a4ad-4b26-8702-a90d9ff156bc"). InnerVolumeSpecName "kube-api-access-99qrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.864605 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.864644 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa6de41-a4ad-4b26-8702-a90d9ff156bc-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.864657 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jfc5\" (UniqueName: \"kubernetes.io/projected/07a57f15-7d0c-468e-9e13-07bf8f493f63-kube-api-access-8jfc5\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.864671 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07a57f15-7d0c-468e-9e13-07bf8f493f63-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:55 crc kubenswrapper[4945]: I1206 07:14:55.864683 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99qrf\" (UniqueName: \"kubernetes.io/projected/faa6de41-a4ad-4b26-8702-a90d9ff156bc-kube-api-access-99qrf\") on node \"crc\" DevicePath \"\"" Dec 06 07:14:56 crc kubenswrapper[4945]: I1206 07:14:56.056104 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-tgskj"] Dec 06 07:14:56 crc kubenswrapper[4945]: I1206 07:14:56.068808 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-tgskj"] Dec 06 07:14:56 crc kubenswrapper[4945]: I1206 07:14:56.091501 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-5bqzf"] Dec 06 07:14:56 crc kubenswrapper[4945]: I1206 07:14:56.099208 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-5bqzf"] Dec 06 07:14:56 crc kubenswrapper[4945]: I1206 07:14:56.965391 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07a57f15-7d0c-468e-9e13-07bf8f493f63" path="/var/lib/kubelet/pods/07a57f15-7d0c-468e-9e13-07bf8f493f63/volumes" Dec 06 07:14:56 crc kubenswrapper[4945]: I1206 07:14:56.965933 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faa6de41-a4ad-4b26-8702-a90d9ff156bc" path="/var/lib/kubelet/pods/faa6de41-a4ad-4b26-8702-a90d9ff156bc/volumes" Dec 06 07:14:57 crc kubenswrapper[4945]: I1206 07:14:57.650666 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-bbc9w"] Dec 06 07:14:57 crc kubenswrapper[4945]: I1206 07:14:57.747596 4945 generic.go:334] "Generic (PLEG): container finished" podID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" containerID="c7eb1e87ecd225ee2282907be972c030861b9852e609a8c7badc9b0c36c11a22" exitCode=0 Dec 06 07:14:57 crc kubenswrapper[4945]: I1206 07:14:57.747729 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ae8dae45-f01e-4ef0-8814-07511ff2fcad","Type":"ContainerDied","Data":"c7eb1e87ecd225ee2282907be972c030861b9852e609a8c7badc9b0c36c11a22"} Dec 06 07:14:57 crc kubenswrapper[4945]: I1206 07:14:57.752073 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" event={"ID":"d175d796-a22c-43e0-af90-739e9dcfb044","Type":"ContainerStarted","Data":"e257b2d1d9144a855034e965f499f07e8c5349888c015e5f42077059a499016e"} Dec 06 07:14:58 crc kubenswrapper[4945]: I1206 07:14:58.087539 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-j4fkv"] Dec 06 07:14:58 crc kubenswrapper[4945]: I1206 07:14:58.135921 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lthjk"] Dec 06 07:14:59 crc kubenswrapper[4945]: I1206 07:14:59.786344 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lthjk" event={"ID":"003bd22c-19f6-4bfa-8f8d-3045dfb7c841","Type":"ContainerStarted","Data":"0327e480f210fd69e40002956bec2d16f8d2d372efb52eb02f84e6de3499153a"} Dec 06 07:14:59 crc kubenswrapper[4945]: I1206 07:14:59.791088 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" event={"ID":"961d0927-7c9d-4b8d-8e5d-a97883f66943","Type":"ContainerStarted","Data":"45cb5c91f502518b701dc7d7ac5dc6ee09340793be8aa51c3ff41003c036cd2e"} Dec 06 07:14:59 crc kubenswrapper[4945]: I1206 07:14:59.795208 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67","Type":"ContainerStarted","Data":"ac89c0e1b702d43f1a10ceb5d1fa331abe2da4bec10b71837cf55f775bb4f9c7"} Dec 06 07:14:59 crc kubenswrapper[4945]: I1206 07:14:59.795297 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 06 07:14:59 crc kubenswrapper[4945]: I1206 07:14:59.802753 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" event={"ID":"d175d796-a22c-43e0-af90-739e9dcfb044","Type":"ContainerStarted","Data":"c7886b7a34f20e0cdf6d6ec6adbafa065e9250a160cb2810f37d7cf572b5f2a6"} Dec 06 07:14:59 crc kubenswrapper[4945]: I1206 07:14:59.804434 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d5496b71-8389-4d16-97c0-5752a2f1ef2a","Type":"ContainerStarted","Data":"ca859b837b4d6d7c6f85efa965868ddba91015b75686b509128eaf6c975919c5"} Dec 06 07:14:59 crc kubenswrapper[4945]: I1206 07:14:59.838083 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=38.677595731 podStartE2EDuration="44.838060259s" podCreationTimestamp="2025-12-06 07:14:15 +0000 UTC" firstStartedPulling="2025-12-06 07:14:51.441861912 +0000 UTC m=+1304.896722956" lastFinishedPulling="2025-12-06 07:14:57.60232644 +0000 UTC m=+1311.057187484" observedRunningTime="2025-12-06 07:14:59.8144433 +0000 UTC m=+1313.269304344" watchObservedRunningTime="2025-12-06 07:14:59.838060259 +0000 UTC m=+1313.292921303" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.183545 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs"] Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.185742 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.188497 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.188845 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.202665 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs"] Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.253591 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/359894c8-e451-4b8c-a0ed-48902377d481-config-volume\") pod \"collect-profiles-29416755-q9gfs\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.253690 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fnt4\" (UniqueName: \"kubernetes.io/projected/359894c8-e451-4b8c-a0ed-48902377d481-kube-api-access-2fnt4\") pod \"collect-profiles-29416755-q9gfs\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.253858 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/359894c8-e451-4b8c-a0ed-48902377d481-secret-volume\") pod \"collect-profiles-29416755-q9gfs\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.355408 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/359894c8-e451-4b8c-a0ed-48902377d481-config-volume\") pod \"collect-profiles-29416755-q9gfs\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.355478 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fnt4\" (UniqueName: \"kubernetes.io/projected/359894c8-e451-4b8c-a0ed-48902377d481-kube-api-access-2fnt4\") pod \"collect-profiles-29416755-q9gfs\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.355603 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/359894c8-e451-4b8c-a0ed-48902377d481-secret-volume\") pod \"collect-profiles-29416755-q9gfs\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.356641 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/359894c8-e451-4b8c-a0ed-48902377d481-config-volume\") pod \"collect-profiles-29416755-q9gfs\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.361825 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/359894c8-e451-4b8c-a0ed-48902377d481-secret-volume\") pod \"collect-profiles-29416755-q9gfs\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.378233 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fnt4\" (UniqueName: \"kubernetes.io/projected/359894c8-e451-4b8c-a0ed-48902377d481-kube-api-access-2fnt4\") pod \"collect-profiles-29416755-q9gfs\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.515901 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.827466 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f5f1ef51-68e5-4133-a248-1d7251b5bb96","Type":"ContainerStarted","Data":"335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16"} Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.827932 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.830858 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xc2dd" event={"ID":"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478","Type":"ContainerStarted","Data":"40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398"} Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.831168 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-xc2dd" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.833893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" event={"ID":"d175d796-a22c-43e0-af90-739e9dcfb044","Type":"ContainerDied","Data":"c7886b7a34f20e0cdf6d6ec6adbafa065e9250a160cb2810f37d7cf572b5f2a6"} Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.834513 4945 generic.go:334] "Generic (PLEG): container finished" podID="d175d796-a22c-43e0-af90-739e9dcfb044" containerID="c7886b7a34f20e0cdf6d6ec6adbafa065e9250a160cb2810f37d7cf572b5f2a6" exitCode=0 Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.853953 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=35.772058968 podStartE2EDuration="43.853928314s" podCreationTimestamp="2025-12-06 07:14:17 +0000 UTC" firstStartedPulling="2025-12-06 07:14:51.351509828 +0000 UTC m=+1304.806370872" lastFinishedPulling="2025-12-06 07:14:59.433379174 +0000 UTC m=+1312.888240218" observedRunningTime="2025-12-06 07:15:00.848516828 +0000 UTC m=+1314.303377882" watchObservedRunningTime="2025-12-06 07:15:00.853928314 +0000 UTC m=+1314.308789358" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.854196 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b51cdd1a-131a-4f4c-8521-93890fd48722","Type":"ContainerStarted","Data":"4441bb3a196d951e00485c2dfb22d064eba695cdb391e51c804ea9be548b968a"} Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.858072 4945 generic.go:334] "Generic (PLEG): container finished" podID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerID="febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92" exitCode=0 Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.858127 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2996k" event={"ID":"9c092be9-37db-49bb-be6b-56ce2e0e15fd","Type":"ContainerDied","Data":"febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92"} Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.861196 4945 generic.go:334] "Generic (PLEG): container finished" podID="961d0927-7c9d-4b8d-8e5d-a97883f66943" containerID="ac8cda00bd3b100a6f760e3738e17ed8b56418887048fa0882387b84b8d0f9cd" exitCode=0 Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.861333 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" event={"ID":"961d0927-7c9d-4b8d-8e5d-a97883f66943","Type":"ContainerDied","Data":"ac8cda00bd3b100a6f760e3738e17ed8b56418887048fa0882387b84b8d0f9cd"} Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.874708 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ae8dae45-f01e-4ef0-8814-07511ff2fcad","Type":"ContainerStarted","Data":"bb3773d5b61628e5effd422bfa3be33b5be1e9ba5f96b66980463e518f4da6ac"} Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.897066 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xc2dd" podStartSLOduration=34.266102227 podStartE2EDuration="40.8970465s" podCreationTimestamp="2025-12-06 07:14:20 +0000 UTC" firstStartedPulling="2025-12-06 07:14:51.329184744 +0000 UTC m=+1304.784045778" lastFinishedPulling="2025-12-06 07:14:57.960129007 +0000 UTC m=+1311.414990051" observedRunningTime="2025-12-06 07:15:00.891339866 +0000 UTC m=+1314.346200930" watchObservedRunningTime="2025-12-06 07:15:00.8970465 +0000 UTC m=+1314.351907544" Dec 06 07:15:00 crc kubenswrapper[4945]: I1206 07:15:00.985558 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=48.381035404 podStartE2EDuration="48.985533064s" podCreationTimestamp="2025-12-06 07:14:12 +0000 UTC" firstStartedPulling="2025-12-06 07:14:51.336541023 +0000 UTC m=+1304.791402067" lastFinishedPulling="2025-12-06 07:14:51.941038683 +0000 UTC m=+1305.395899727" observedRunningTime="2025-12-06 07:15:00.981684629 +0000 UTC m=+1314.436545673" watchObservedRunningTime="2025-12-06 07:15:00.985533064 +0000 UTC m=+1314.440394128" Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.110722 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs"] Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.886599 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" event={"ID":"961d0927-7c9d-4b8d-8e5d-a97883f66943","Type":"ContainerStarted","Data":"1d6bb924dabf092bdb00ec143c80ce94adeb73b6e6f6f6374bb91fe10413131c"} Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.886884 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.888578 4945 generic.go:334] "Generic (PLEG): container finished" podID="359894c8-e451-4b8c-a0ed-48902377d481" containerID="640e52dff2a20bc3c593474e6eb40ae052b77fc371b61a009179e281519d3245" exitCode=0 Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.888629 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" event={"ID":"359894c8-e451-4b8c-a0ed-48902377d481","Type":"ContainerDied","Data":"640e52dff2a20bc3c593474e6eb40ae052b77fc371b61a009179e281519d3245"} Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.888649 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" event={"ID":"359894c8-e451-4b8c-a0ed-48902377d481","Type":"ContainerStarted","Data":"e67656670bb3b13bf5deb36f4b56787d2c1d490309505ea71de2a95e3f76e755"} Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.891433 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" event={"ID":"d175d796-a22c-43e0-af90-739e9dcfb044","Type":"ContainerStarted","Data":"6bd1a2caf220155e80ac2225c1daa30340dfa2008bdbfbddacd2c550fef48aac"} Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.891584 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.893923 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2996k" event={"ID":"9c092be9-37db-49bb-be6b-56ce2e0e15fd","Type":"ContainerStarted","Data":"59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600"} Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.893948 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2996k" event={"ID":"9c092be9-37db-49bb-be6b-56ce2e0e15fd","Type":"ContainerStarted","Data":"ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e"} Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.896613 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.897407 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.922657 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" podStartSLOduration=8.485700401999999 podStartE2EDuration="8.922634249s" podCreationTimestamp="2025-12-06 07:14:53 +0000 UTC" firstStartedPulling="2025-12-06 07:14:59.394335018 +0000 UTC m=+1312.849196062" lastFinishedPulling="2025-12-06 07:14:59.831268855 +0000 UTC m=+1313.286129909" observedRunningTime="2025-12-06 07:15:01.917779248 +0000 UTC m=+1315.372640312" watchObservedRunningTime="2025-12-06 07:15:01.922634249 +0000 UTC m=+1315.377495293" Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.966118 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" podStartSLOduration=7.184670332 podStartE2EDuration="8.966097424s" podCreationTimestamp="2025-12-06 07:14:53 +0000 UTC" firstStartedPulling="2025-12-06 07:14:57.651471169 +0000 UTC m=+1311.106332213" lastFinishedPulling="2025-12-06 07:14:59.432898261 +0000 UTC m=+1312.887759305" observedRunningTime="2025-12-06 07:15:01.962578269 +0000 UTC m=+1315.417439343" watchObservedRunningTime="2025-12-06 07:15:01.966097424 +0000 UTC m=+1315.420958468" Dec 06 07:15:01 crc kubenswrapper[4945]: I1206 07:15:01.996953 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-2996k" podStartSLOduration=35.096659841 podStartE2EDuration="41.996923668s" podCreationTimestamp="2025-12-06 07:14:20 +0000 UTC" firstStartedPulling="2025-12-06 07:14:50.702132855 +0000 UTC m=+1304.156993889" lastFinishedPulling="2025-12-06 07:14:57.602396662 +0000 UTC m=+1311.057257716" observedRunningTime="2025-12-06 07:15:01.990125204 +0000 UTC m=+1315.444986258" watchObservedRunningTime="2025-12-06 07:15:01.996923668 +0000 UTC m=+1315.451784712" Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.636106 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:03 crc kubenswrapper[4945]: E1206 07:15:03.640005 4945 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.220:37000->38.129.56.220:39753: read tcp 38.129.56.220:37000->38.129.56.220:39753: read: connection reset by peer Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.753568 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fnt4\" (UniqueName: \"kubernetes.io/projected/359894c8-e451-4b8c-a0ed-48902377d481-kube-api-access-2fnt4\") pod \"359894c8-e451-4b8c-a0ed-48902377d481\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.753769 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/359894c8-e451-4b8c-a0ed-48902377d481-secret-volume\") pod \"359894c8-e451-4b8c-a0ed-48902377d481\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.753821 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/359894c8-e451-4b8c-a0ed-48902377d481-config-volume\") pod \"359894c8-e451-4b8c-a0ed-48902377d481\" (UID: \"359894c8-e451-4b8c-a0ed-48902377d481\") " Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.755123 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/359894c8-e451-4b8c-a0ed-48902377d481-config-volume" (OuterVolumeSpecName: "config-volume") pod "359894c8-e451-4b8c-a0ed-48902377d481" (UID: "359894c8-e451-4b8c-a0ed-48902377d481"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.765065 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/359894c8-e451-4b8c-a0ed-48902377d481-kube-api-access-2fnt4" (OuterVolumeSpecName: "kube-api-access-2fnt4") pod "359894c8-e451-4b8c-a0ed-48902377d481" (UID: "359894c8-e451-4b8c-a0ed-48902377d481"). InnerVolumeSpecName "kube-api-access-2fnt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.766008 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/359894c8-e451-4b8c-a0ed-48902377d481-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "359894c8-e451-4b8c-a0ed-48902377d481" (UID: "359894c8-e451-4b8c-a0ed-48902377d481"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.836107 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.836159 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.855846 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/359894c8-e451-4b8c-a0ed-48902377d481-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.855883 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/359894c8-e451-4b8c-a0ed-48902377d481-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.855894 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fnt4\" (UniqueName: \"kubernetes.io/projected/359894c8-e451-4b8c-a0ed-48902377d481-kube-api-access-2fnt4\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.921247 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" event={"ID":"359894c8-e451-4b8c-a0ed-48902377d481","Type":"ContainerDied","Data":"e67656670bb3b13bf5deb36f4b56787d2c1d490309505ea71de2a95e3f76e755"} Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.921302 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e67656670bb3b13bf5deb36f4b56787d2c1d490309505ea71de2a95e3f76e755" Dec 06 07:15:03 crc kubenswrapper[4945]: I1206 07:15:03.921340 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs" Dec 06 07:15:04 crc kubenswrapper[4945]: I1206 07:15:04.931582 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fe30cfa7-9746-4771-bac5-29e3a3662aec","Type":"ContainerStarted","Data":"9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b"} Dec 06 07:15:04 crc kubenswrapper[4945]: I1206 07:15:04.936501 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d5496b71-8389-4d16-97c0-5752a2f1ef2a","Type":"ContainerStarted","Data":"8a0182213333baacc78f792add7c9120ea279ab7a0ddd78d4a1493c3debaba17"} Dec 06 07:15:04 crc kubenswrapper[4945]: I1206 07:15:04.940022 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b51cdd1a-131a-4f4c-8521-93890fd48722","Type":"ContainerStarted","Data":"dcc267ea7f865def7c69c6af3584068ad5567448a3b0f26d54246303ab0aeed6"} Dec 06 07:15:04 crc kubenswrapper[4945]: I1206 07:15:04.942437 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lthjk" event={"ID":"003bd22c-19f6-4bfa-8f8d-3045dfb7c841","Type":"ContainerStarted","Data":"7e2f33a76160a89b2b209c529e605d89f63fe75180ef0858d4423cb32db2c20d"} Dec 06 07:15:05 crc kubenswrapper[4945]: I1206 07:15:05.040142 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=35.122515771 podStartE2EDuration="48.040119265s" podCreationTimestamp="2025-12-06 07:14:17 +0000 UTC" firstStartedPulling="2025-12-06 07:14:51.52903504 +0000 UTC m=+1304.983896094" lastFinishedPulling="2025-12-06 07:15:04.446638544 +0000 UTC m=+1317.901499588" observedRunningTime="2025-12-06 07:15:05.034078092 +0000 UTC m=+1318.488939156" watchObservedRunningTime="2025-12-06 07:15:05.040119265 +0000 UTC m=+1318.494980309" Dec 06 07:15:05 crc kubenswrapper[4945]: I1206 07:15:05.061888 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-lthjk" podStartSLOduration=6.89319705 podStartE2EDuration="12.061864923s" podCreationTimestamp="2025-12-06 07:14:53 +0000 UTC" firstStartedPulling="2025-12-06 07:14:59.33080344 +0000 UTC m=+1312.785664484" lastFinishedPulling="2025-12-06 07:15:04.499471313 +0000 UTC m=+1317.954332357" observedRunningTime="2025-12-06 07:15:05.054970037 +0000 UTC m=+1318.509831081" watchObservedRunningTime="2025-12-06 07:15:05.061864923 +0000 UTC m=+1318.516725967" Dec 06 07:15:05 crc kubenswrapper[4945]: I1206 07:15:05.082498 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=31.947521855 podStartE2EDuration="44.082478941s" podCreationTimestamp="2025-12-06 07:14:21 +0000 UTC" firstStartedPulling="2025-12-06 07:14:52.343259912 +0000 UTC m=+1305.798120956" lastFinishedPulling="2025-12-06 07:15:04.478216998 +0000 UTC m=+1317.933078042" observedRunningTime="2025-12-06 07:15:05.080562239 +0000 UTC m=+1318.535423283" watchObservedRunningTime="2025-12-06 07:15:05.082478941 +0000 UTC m=+1318.537339985" Dec 06 07:15:05 crc kubenswrapper[4945]: I1206 07:15:05.275514 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 06 07:15:05 crc kubenswrapper[4945]: I1206 07:15:05.316094 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 06 07:15:05 crc kubenswrapper[4945]: I1206 07:15:05.815428 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 06 07:15:05 crc kubenswrapper[4945]: I1206 07:15:05.951034 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 06 07:15:05 crc kubenswrapper[4945]: I1206 07:15:05.997053 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 06 07:15:06 crc kubenswrapper[4945]: I1206 07:15:06.185764 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 06 07:15:06 crc kubenswrapper[4945]: I1206 07:15:06.286155 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.326306 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.374789 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.630428 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.749437 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-bbc9w"] Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.749747 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" podUID="d175d796-a22c-43e0-af90-739e9dcfb044" containerName="dnsmasq-dns" containerID="cri-o://6bd1a2caf220155e80ac2225c1daa30340dfa2008bdbfbddacd2c550fef48aac" gracePeriod=10 Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.750415 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.783318 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-mlclv"] Dec 06 07:15:07 crc kubenswrapper[4945]: E1206 07:15:07.784091 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="359894c8-e451-4b8c-a0ed-48902377d481" containerName="collect-profiles" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.784115 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="359894c8-e451-4b8c-a0ed-48902377d481" containerName="collect-profiles" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.784370 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="359894c8-e451-4b8c-a0ed-48902377d481" containerName="collect-profiles" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.785622 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.830615 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-mlclv"] Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.952233 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.952332 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-config\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.952376 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-dns-svc\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.952430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.952500 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwbtj\" (UniqueName: \"kubernetes.io/projected/0f520e1c-bdfd-4fc9-9608-237d44a73622-kube-api-access-bwbtj\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.972296 4945 generic.go:334] "Generic (PLEG): container finished" podID="d175d796-a22c-43e0-af90-739e9dcfb044" containerID="6bd1a2caf220155e80ac2225c1daa30340dfa2008bdbfbddacd2c550fef48aac" exitCode=0 Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.973172 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" event={"ID":"d175d796-a22c-43e0-af90-739e9dcfb044","Type":"ContainerDied","Data":"6bd1a2caf220155e80ac2225c1daa30340dfa2008bdbfbddacd2c550fef48aac"} Dec 06 07:15:07 crc kubenswrapper[4945]: I1206 07:15:07.973214 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.043105 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.054258 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.054372 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-config\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.054433 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-dns-svc\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.054524 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.054611 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwbtj\" (UniqueName: \"kubernetes.io/projected/0f520e1c-bdfd-4fc9-9608-237d44a73622-kube-api-access-bwbtj\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.056265 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.056588 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.063089 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-dns-svc\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.066090 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-config\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.102974 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwbtj\" (UniqueName: \"kubernetes.io/projected/0f520e1c-bdfd-4fc9-9608-237d44a73622-kube-api-access-bwbtj\") pod \"dnsmasq-dns-59d5fbdd8c-mlclv\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.129538 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.236329 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.238039 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.246536 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.265903 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-n7j6n" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.266138 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.266315 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.285991 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.360630 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-scripts\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.360745 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.360819 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-config\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.360890 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnrx2\" (UniqueName: \"kubernetes.io/projected/232baa8c-2d18-49f3-b230-362181e687df-kube-api-access-lnrx2\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.361003 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.361213 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.361309 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/232baa8c-2d18-49f3-b230-362181e687df-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.411876 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.463021 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnrx2\" (UniqueName: \"kubernetes.io/projected/232baa8c-2d18-49f3-b230-362181e687df-kube-api-access-lnrx2\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.463072 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.463097 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.463141 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/232baa8c-2d18-49f3-b230-362181e687df-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.463196 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-scripts\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.463869 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.464010 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-config\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.464579 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/232baa8c-2d18-49f3-b230-362181e687df-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.464585 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-scripts\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.465127 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-config\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.471731 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.472299 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.472412 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.496586 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnrx2\" (UniqueName: \"kubernetes.io/projected/232baa8c-2d18-49f3-b230-362181e687df-kube-api-access-lnrx2\") pod \"ovn-northd-0\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.564983 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k948d\" (UniqueName: \"kubernetes.io/projected/d175d796-a22c-43e0-af90-739e9dcfb044-kube-api-access-k948d\") pod \"d175d796-a22c-43e0-af90-739e9dcfb044\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.565125 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-dns-svc\") pod \"d175d796-a22c-43e0-af90-739e9dcfb044\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.565178 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-ovsdbserver-nb\") pod \"d175d796-a22c-43e0-af90-739e9dcfb044\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.565298 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-config\") pod \"d175d796-a22c-43e0-af90-739e9dcfb044\" (UID: \"d175d796-a22c-43e0-af90-739e9dcfb044\") " Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.570666 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d175d796-a22c-43e0-af90-739e9dcfb044-kube-api-access-k948d" (OuterVolumeSpecName: "kube-api-access-k948d") pod "d175d796-a22c-43e0-af90-739e9dcfb044" (UID: "d175d796-a22c-43e0-af90-739e9dcfb044"). InnerVolumeSpecName "kube-api-access-k948d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.620519 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d175d796-a22c-43e0-af90-739e9dcfb044" (UID: "d175d796-a22c-43e0-af90-739e9dcfb044"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.621097 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.636670 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-config" (OuterVolumeSpecName: "config") pod "d175d796-a22c-43e0-af90-739e9dcfb044" (UID: "d175d796-a22c-43e0-af90-739e9dcfb044"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.647424 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d175d796-a22c-43e0-af90-739e9dcfb044" (UID: "d175d796-a22c-43e0-af90-739e9dcfb044"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.669898 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k948d\" (UniqueName: \"kubernetes.io/projected/d175d796-a22c-43e0-af90-739e9dcfb044-kube-api-access-k948d\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.669941 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.669955 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.669963 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d175d796-a22c-43e0-af90-739e9dcfb044-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.803178 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-mlclv"] Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.952480 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.978895 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 06 07:15:08 crc kubenswrapper[4945]: E1206 07:15:08.979346 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d175d796-a22c-43e0-af90-739e9dcfb044" containerName="init" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.979362 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d175d796-a22c-43e0-af90-739e9dcfb044" containerName="init" Dec 06 07:15:08 crc kubenswrapper[4945]: E1206 07:15:08.979381 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d175d796-a22c-43e0-af90-739e9dcfb044" containerName="dnsmasq-dns" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.979387 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d175d796-a22c-43e0-af90-739e9dcfb044" containerName="dnsmasq-dns" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.980145 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d175d796-a22c-43e0-af90-739e9dcfb044" containerName="dnsmasq-dns" Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.996843 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 07:15:08 crc kubenswrapper[4945]: I1206 07:15:08.998101 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.001012 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"232baa8c-2d18-49f3-b230-362181e687df","Type":"ContainerStarted","Data":"0d2515fb2c0ed415c60853fb62ac0440c1b45493e80440f481ade0b520891668"} Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.001364 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.002763 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.002922 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.003122 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-mslnb" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.008998 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" event={"ID":"d175d796-a22c-43e0-af90-739e9dcfb044","Type":"ContainerDied","Data":"e257b2d1d9144a855034e965f499f07e8c5349888c015e5f42077059a499016e"} Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.009045 4945 scope.go:117] "RemoveContainer" containerID="6bd1a2caf220155e80ac2225c1daa30340dfa2008bdbfbddacd2c550fef48aac" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.009164 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-bbc9w" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.015777 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" event={"ID":"0f520e1c-bdfd-4fc9-9608-237d44a73622","Type":"ContainerStarted","Data":"871f87cc86619393b4f8c2501f9744ca3c69272b990f0c4ae419fe0df291d459"} Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.049111 4945 scope.go:117] "RemoveContainer" containerID="c7886b7a34f20e0cdf6d6ec6adbafa065e9250a160cb2810f37d7cf572b5f2a6" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.069540 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-bbc9w"] Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.077080 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-bbc9w"] Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.079914 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-lock\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.079976 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v99kc\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-kube-api-access-v99kc\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.080024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.080119 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-cache\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.080143 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.143466 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.181556 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-lock\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.181611 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v99kc\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-kube-api-access-v99kc\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.181651 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.181668 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-cache\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.181684 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: E1206 07:15:09.181929 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 07:15:09 crc kubenswrapper[4945]: E1206 07:15:09.181945 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 07:15:09 crc kubenswrapper[4945]: E1206 07:15:09.181992 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift podName:af58797c-3af9-4731-88dc-5c03e52217b4 nodeName:}" failed. No retries permitted until 2025-12-06 07:15:09.681974578 +0000 UTC m=+1323.136835622 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift") pod "swift-storage-0" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4") : configmap "swift-ring-files" not found Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.182505 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-lock\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.182738 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-cache\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.182802 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.219696 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.221025 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v99kc\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-kube-api-access-v99kc\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: I1206 07:15:09.690403 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:09 crc kubenswrapper[4945]: E1206 07:15:09.690586 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 07:15:09 crc kubenswrapper[4945]: E1206 07:15:09.690600 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 07:15:09 crc kubenswrapper[4945]: E1206 07:15:09.690652 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift podName:af58797c-3af9-4731-88dc-5c03e52217b4 nodeName:}" failed. No retries permitted until 2025-12-06 07:15:10.690637455 +0000 UTC m=+1324.145498499 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift") pod "swift-storage-0" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4") : configmap "swift-ring-files" not found Dec 06 07:15:10 crc kubenswrapper[4945]: I1206 07:15:10.708322 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:10 crc kubenswrapper[4945]: E1206 07:15:10.708590 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 07:15:10 crc kubenswrapper[4945]: E1206 07:15:10.709534 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 07:15:10 crc kubenswrapper[4945]: E1206 07:15:10.709623 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift podName:af58797c-3af9-4731-88dc-5c03e52217b4 nodeName:}" failed. No retries permitted until 2025-12-06 07:15:12.709600055 +0000 UTC m=+1326.164461109 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift") pod "swift-storage-0" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4") : configmap "swift-ring-files" not found Dec 06 07:15:10 crc kubenswrapper[4945]: I1206 07:15:10.966387 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d175d796-a22c-43e0-af90-739e9dcfb044" path="/var/lib/kubelet/pods/d175d796-a22c-43e0-af90-739e9dcfb044/volumes" Dec 06 07:15:11 crc kubenswrapper[4945]: I1206 07:15:11.039728 4945 generic.go:334] "Generic (PLEG): container finished" podID="fe30cfa7-9746-4771-bac5-29e3a3662aec" containerID="9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b" exitCode=0 Dec 06 07:15:11 crc kubenswrapper[4945]: I1206 07:15:11.039907 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fe30cfa7-9746-4771-bac5-29e3a3662aec","Type":"ContainerDied","Data":"9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b"} Dec 06 07:15:11 crc kubenswrapper[4945]: I1206 07:15:11.042778 4945 generic.go:334] "Generic (PLEG): container finished" podID="0f520e1c-bdfd-4fc9-9608-237d44a73622" containerID="50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6" exitCode=0 Dec 06 07:15:11 crc kubenswrapper[4945]: I1206 07:15:11.042902 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" event={"ID":"0f520e1c-bdfd-4fc9-9608-237d44a73622","Type":"ContainerDied","Data":"50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6"} Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.061173 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fe30cfa7-9746-4771-bac5-29e3a3662aec","Type":"ContainerStarted","Data":"bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8"} Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.064467 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" event={"ID":"0f520e1c-bdfd-4fc9-9608-237d44a73622","Type":"ContainerStarted","Data":"877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937"} Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.066918 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"232baa8c-2d18-49f3-b230-362181e687df","Type":"ContainerStarted","Data":"db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba"} Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.066973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"232baa8c-2d18-49f3-b230-362181e687df","Type":"ContainerStarted","Data":"01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4"} Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.067206 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.086069 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371975.768734 podStartE2EDuration="1m1.086041242s" podCreationTimestamp="2025-12-06 07:14:11 +0000 UTC" firstStartedPulling="2025-12-06 07:14:13.267478893 +0000 UTC m=+1266.722339937" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:12.081508609 +0000 UTC m=+1325.536369673" watchObservedRunningTime="2025-12-06 07:15:12.086041242 +0000 UTC m=+1325.540902286" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.112009 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.085568566 podStartE2EDuration="4.111986224s" podCreationTimestamp="2025-12-06 07:15:08 +0000 UTC" firstStartedPulling="2025-12-06 07:15:08.967588789 +0000 UTC m=+1322.422449843" lastFinishedPulling="2025-12-06 07:15:10.994006457 +0000 UTC m=+1324.448867501" observedRunningTime="2025-12-06 07:15:12.102884408 +0000 UTC m=+1325.557745472" watchObservedRunningTime="2025-12-06 07:15:12.111986224 +0000 UTC m=+1325.566847268" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.127857 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" podStartSLOduration=5.127839092 podStartE2EDuration="5.127839092s" podCreationTimestamp="2025-12-06 07:15:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:12.122733974 +0000 UTC m=+1325.577595038" watchObservedRunningTime="2025-12-06 07:15:12.127839092 +0000 UTC m=+1325.582700136" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.582089 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.582354 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.750264 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:12 crc kubenswrapper[4945]: E1206 07:15:12.750605 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 07:15:12 crc kubenswrapper[4945]: E1206 07:15:12.750628 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 07:15:12 crc kubenswrapper[4945]: E1206 07:15:12.750685 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift podName:af58797c-3af9-4731-88dc-5c03e52217b4 nodeName:}" failed. No retries permitted until 2025-12-06 07:15:16.750665588 +0000 UTC m=+1330.205526632 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift") pod "swift-storage-0" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4") : configmap "swift-ring-files" not found Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.820646 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-hjh59"] Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.821872 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.824147 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.824709 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.825014 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.831016 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hjh59"] Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.851984 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3603082b-401e-484c-88a3-90c8f67f5824-etc-swift\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.852029 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-dispersionconf\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.852106 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-scripts\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.852143 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktmfs\" (UniqueName: \"kubernetes.io/projected/3603082b-401e-484c-88a3-90c8f67f5824-kube-api-access-ktmfs\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.852248 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-swiftconf\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.852273 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-ring-data-devices\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.852324 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-combined-ca-bundle\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.953954 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-swiftconf\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.955160 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-ring-data-devices\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.955216 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-combined-ca-bundle\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.955241 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3603082b-401e-484c-88a3-90c8f67f5824-etc-swift\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.955262 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-dispersionconf\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.955525 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-scripts\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.956068 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3603082b-401e-484c-88a3-90c8f67f5824-etc-swift\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.956095 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-ring-data-devices\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.956181 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-scripts\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.956411 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktmfs\" (UniqueName: \"kubernetes.io/projected/3603082b-401e-484c-88a3-90c8f67f5824-kube-api-access-ktmfs\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.959902 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-dispersionconf\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.962763 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-swiftconf\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.973039 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-combined-ca-bundle\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:12 crc kubenswrapper[4945]: I1206 07:15:12.979549 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktmfs\" (UniqueName: \"kubernetes.io/projected/3603082b-401e-484c-88a3-90c8f67f5824-kube-api-access-ktmfs\") pod \"swift-ring-rebalance-hjh59\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:13 crc kubenswrapper[4945]: I1206 07:15:13.073581 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:13 crc kubenswrapper[4945]: I1206 07:15:13.151146 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:13 crc kubenswrapper[4945]: I1206 07:15:13.598554 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hjh59"] Dec 06 07:15:14 crc kubenswrapper[4945]: I1206 07:15:14.081619 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hjh59" event={"ID":"3603082b-401e-484c-88a3-90c8f67f5824","Type":"ContainerStarted","Data":"8ed5356b9ec4b2f75d7097d794e6ab0a9d18c0f8ea781e3b6d63644e2953f8cd"} Dec 06 07:15:14 crc kubenswrapper[4945]: E1206 07:15:14.483486 4945 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.220:56468->38.129.56.220:39753: write tcp 38.129.56.220:56468->38.129.56.220:39753: write: connection reset by peer Dec 06 07:15:16 crc kubenswrapper[4945]: I1206 07:15:16.675619 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 06 07:15:16 crc kubenswrapper[4945]: I1206 07:15:16.754592 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 06 07:15:16 crc kubenswrapper[4945]: I1206 07:15:16.832569 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:16 crc kubenswrapper[4945]: E1206 07:15:16.842522 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 07:15:16 crc kubenswrapper[4945]: E1206 07:15:16.842561 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 07:15:16 crc kubenswrapper[4945]: E1206 07:15:16.842638 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift podName:af58797c-3af9-4731-88dc-5c03e52217b4 nodeName:}" failed. No retries permitted until 2025-12-06 07:15:24.842615011 +0000 UTC m=+1338.297476055 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift") pod "swift-storage-0" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4") : configmap "swift-ring-files" not found Dec 06 07:15:18 crc kubenswrapper[4945]: I1206 07:15:18.131777 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:15:18 crc kubenswrapper[4945]: I1206 07:15:18.208430 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-j4fkv"] Dec 06 07:15:18 crc kubenswrapper[4945]: I1206 07:15:18.208928 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" podUID="961d0927-7c9d-4b8d-8e5d-a97883f66943" containerName="dnsmasq-dns" containerID="cri-o://1d6bb924dabf092bdb00ec143c80ce94adeb73b6e6f6f6374bb91fe10413131c" gracePeriod=10 Dec 06 07:15:19 crc kubenswrapper[4945]: I1206 07:15:19.123192 4945 generic.go:334] "Generic (PLEG): container finished" podID="961d0927-7c9d-4b8d-8e5d-a97883f66943" containerID="1d6bb924dabf092bdb00ec143c80ce94adeb73b6e6f6f6374bb91fe10413131c" exitCode=0 Dec 06 07:15:19 crc kubenswrapper[4945]: I1206 07:15:19.123542 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" event={"ID":"961d0927-7c9d-4b8d-8e5d-a97883f66943","Type":"ContainerDied","Data":"1d6bb924dabf092bdb00ec143c80ce94adeb73b6e6f6f6374bb91fe10413131c"} Dec 06 07:15:19 crc kubenswrapper[4945]: I1206 07:15:19.219009 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" podUID="961d0927-7c9d-4b8d-8e5d-a97883f66943" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.503070 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.666801 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-config\") pod \"961d0927-7c9d-4b8d-8e5d-a97883f66943\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.667016 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnkbg\" (UniqueName: \"kubernetes.io/projected/961d0927-7c9d-4b8d-8e5d-a97883f66943-kube-api-access-pnkbg\") pod \"961d0927-7c9d-4b8d-8e5d-a97883f66943\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.667059 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-dns-svc\") pod \"961d0927-7c9d-4b8d-8e5d-a97883f66943\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.667149 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-sb\") pod \"961d0927-7c9d-4b8d-8e5d-a97883f66943\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.667213 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-nb\") pod \"961d0927-7c9d-4b8d-8e5d-a97883f66943\" (UID: \"961d0927-7c9d-4b8d-8e5d-a97883f66943\") " Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.673892 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/961d0927-7c9d-4b8d-8e5d-a97883f66943-kube-api-access-pnkbg" (OuterVolumeSpecName: "kube-api-access-pnkbg") pod "961d0927-7c9d-4b8d-8e5d-a97883f66943" (UID: "961d0927-7c9d-4b8d-8e5d-a97883f66943"). InnerVolumeSpecName "kube-api-access-pnkbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.687226 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.708144 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "961d0927-7c9d-4b8d-8e5d-a97883f66943" (UID: "961d0927-7c9d-4b8d-8e5d-a97883f66943"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.718081 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "961d0927-7c9d-4b8d-8e5d-a97883f66943" (UID: "961d0927-7c9d-4b8d-8e5d-a97883f66943"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.718157 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "961d0927-7c9d-4b8d-8e5d-a97883f66943" (UID: "961d0927-7c9d-4b8d-8e5d-a97883f66943"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.718870 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-config" (OuterVolumeSpecName: "config") pod "961d0927-7c9d-4b8d-8e5d-a97883f66943" (UID: "961d0927-7c9d-4b8d-8e5d-a97883f66943"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.770000 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.770032 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnkbg\" (UniqueName: \"kubernetes.io/projected/961d0927-7c9d-4b8d-8e5d-a97883f66943-kube-api-access-pnkbg\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.770045 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.770054 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:23 crc kubenswrapper[4945]: I1206 07:15:23.770063 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/961d0927-7c9d-4b8d-8e5d-a97883f66943-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.185805 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" event={"ID":"961d0927-7c9d-4b8d-8e5d-a97883f66943","Type":"ContainerDied","Data":"45cb5c91f502518b701dc7d7ac5dc6ee09340793be8aa51c3ff41003c036cd2e"} Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.185832 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-j4fkv" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.185875 4945 scope.go:117] "RemoveContainer" containerID="1d6bb924dabf092bdb00ec143c80ce94adeb73b6e6f6f6374bb91fe10413131c" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.189497 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hjh59" event={"ID":"3603082b-401e-484c-88a3-90c8f67f5824","Type":"ContainerStarted","Data":"3624afb9c68cb040abd682f4ced6ee22803012bbc4d7d8b26dd3e7ce553edb4c"} Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.207315 4945 scope.go:117] "RemoveContainer" containerID="ac8cda00bd3b100a6f760e3738e17ed8b56418887048fa0882387b84b8d0f9cd" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.224774 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-hjh59" podStartSLOduration=3.018403183 podStartE2EDuration="12.224747451s" podCreationTimestamp="2025-12-06 07:15:12 +0000 UTC" firstStartedPulling="2025-12-06 07:15:13.603817603 +0000 UTC m=+1327.058678647" lastFinishedPulling="2025-12-06 07:15:22.810161871 +0000 UTC m=+1336.265022915" observedRunningTime="2025-12-06 07:15:24.220127156 +0000 UTC m=+1337.674988200" watchObservedRunningTime="2025-12-06 07:15:24.224747451 +0000 UTC m=+1337.679608495" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.242585 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-m5ptv"] Dec 06 07:15:24 crc kubenswrapper[4945]: E1206 07:15:24.243068 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="961d0927-7c9d-4b8d-8e5d-a97883f66943" containerName="init" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.243086 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="961d0927-7c9d-4b8d-8e5d-a97883f66943" containerName="init" Dec 06 07:15:24 crc kubenswrapper[4945]: E1206 07:15:24.243106 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="961d0927-7c9d-4b8d-8e5d-a97883f66943" containerName="dnsmasq-dns" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.243112 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="961d0927-7c9d-4b8d-8e5d-a97883f66943" containerName="dnsmasq-dns" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.243413 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="961d0927-7c9d-4b8d-8e5d-a97883f66943" containerName="dnsmasq-dns" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.244043 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m5ptv" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.250534 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-j4fkv"] Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.258044 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-j4fkv"] Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.275579 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-m5ptv"] Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.386470 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e2897e5-a177-4cb5-9790-4fa9bbb35104-operator-scripts\") pod \"keystone-db-create-m5ptv\" (UID: \"5e2897e5-a177-4cb5-9790-4fa9bbb35104\") " pod="openstack/keystone-db-create-m5ptv" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.386629 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqzrm\" (UniqueName: \"kubernetes.io/projected/5e2897e5-a177-4cb5-9790-4fa9bbb35104-kube-api-access-bqzrm\") pod \"keystone-db-create-m5ptv\" (UID: \"5e2897e5-a177-4cb5-9790-4fa9bbb35104\") " pod="openstack/keystone-db-create-m5ptv" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.487679 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e2897e5-a177-4cb5-9790-4fa9bbb35104-operator-scripts\") pod \"keystone-db-create-m5ptv\" (UID: \"5e2897e5-a177-4cb5-9790-4fa9bbb35104\") " pod="openstack/keystone-db-create-m5ptv" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.487761 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqzrm\" (UniqueName: \"kubernetes.io/projected/5e2897e5-a177-4cb5-9790-4fa9bbb35104-kube-api-access-bqzrm\") pod \"keystone-db-create-m5ptv\" (UID: \"5e2897e5-a177-4cb5-9790-4fa9bbb35104\") " pod="openstack/keystone-db-create-m5ptv" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.488835 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e2897e5-a177-4cb5-9790-4fa9bbb35104-operator-scripts\") pod \"keystone-db-create-m5ptv\" (UID: \"5e2897e5-a177-4cb5-9790-4fa9bbb35104\") " pod="openstack/keystone-db-create-m5ptv" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.514309 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqzrm\" (UniqueName: \"kubernetes.io/projected/5e2897e5-a177-4cb5-9790-4fa9bbb35104-kube-api-access-bqzrm\") pod \"keystone-db-create-m5ptv\" (UID: \"5e2897e5-a177-4cb5-9790-4fa9bbb35104\") " pod="openstack/keystone-db-create-m5ptv" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.574578 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m5ptv" Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.895247 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:24 crc kubenswrapper[4945]: E1206 07:15:24.895469 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 07:15:24 crc kubenswrapper[4945]: E1206 07:15:24.895713 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 07:15:24 crc kubenswrapper[4945]: E1206 07:15:24.895788 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift podName:af58797c-3af9-4731-88dc-5c03e52217b4 nodeName:}" failed. No retries permitted until 2025-12-06 07:15:40.89576681 +0000 UTC m=+1354.350627854 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift") pod "swift-storage-0" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4") : configmap "swift-ring-files" not found Dec 06 07:15:24 crc kubenswrapper[4945]: I1206 07:15:24.964538 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="961d0927-7c9d-4b8d-8e5d-a97883f66943" path="/var/lib/kubelet/pods/961d0927-7c9d-4b8d-8e5d-a97883f66943/volumes" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.180233 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-m5ptv"] Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.218064 4945 generic.go:334] "Generic (PLEG): container finished" podID="250c85e2-e14d-4f93-822c-c9a88352eafe" containerID="3da187010c6d527c479f778bf05d2b504b2c0f1fcf3e34a35517eca9354e58fd" exitCode=0 Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.218349 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"250c85e2-e14d-4f93-822c-c9a88352eafe","Type":"ContainerDied","Data":"3da187010c6d527c479f778bf05d2b504b2c0f1fcf3e34a35517eca9354e58fd"} Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.226420 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-fa31-account-create-update-dbjwp"] Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.234553 4945 generic.go:334] "Generic (PLEG): container finished" podID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" containerID="299978e81ebf6ee60eff482d48ee097c771fe6fe34e8dc8d956786247712867f" exitCode=0 Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.246543 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fa31-account-create-update-dbjwp"] Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.246590 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-m5ptv" event={"ID":"5e2897e5-a177-4cb5-9790-4fa9bbb35104","Type":"ContainerStarted","Data":"6182a1bbfd1ca2dcd498015cec1d5509c804d28281c1e46ec96a2d57a160b99a"} Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.246615 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d71d0084-b4bc-4be5-971b-eced6e6d66ea","Type":"ContainerDied","Data":"299978e81ebf6ee60eff482d48ee097c771fe6fe34e8dc8d956786247712867f"} Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.247845 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fa31-account-create-update-dbjwp" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.250655 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.325128 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-ljfg7"] Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.326942 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ljfg7" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.355510 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ljfg7"] Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.413944 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18f99e6e-1055-4fef-b134-e49c03a0e06a-operator-scripts\") pod \"keystone-fa31-account-create-update-dbjwp\" (UID: \"18f99e6e-1055-4fef-b134-e49c03a0e06a\") " pod="openstack/keystone-fa31-account-create-update-dbjwp" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.419078 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w8q2\" (UniqueName: \"kubernetes.io/projected/18f99e6e-1055-4fef-b134-e49c03a0e06a-kube-api-access-6w8q2\") pod \"keystone-fa31-account-create-update-dbjwp\" (UID: \"18f99e6e-1055-4fef-b134-e49c03a0e06a\") " pod="openstack/keystone-fa31-account-create-update-dbjwp" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.427587 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-73b1-account-create-update-8sddg"] Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.428913 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-73b1-account-create-update-8sddg" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.432836 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.439056 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-73b1-account-create-update-8sddg"] Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.522121 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r7lz\" (UniqueName: \"kubernetes.io/projected/35daefb6-843e-4791-8ef8-c9f850b7bf43-kube-api-access-6r7lz\") pod \"placement-db-create-ljfg7\" (UID: \"35daefb6-843e-4791-8ef8-c9f850b7bf43\") " pod="openstack/placement-db-create-ljfg7" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.522181 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18f99e6e-1055-4fef-b134-e49c03a0e06a-operator-scripts\") pod \"keystone-fa31-account-create-update-dbjwp\" (UID: \"18f99e6e-1055-4fef-b134-e49c03a0e06a\") " pod="openstack/keystone-fa31-account-create-update-dbjwp" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.522228 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w8q2\" (UniqueName: \"kubernetes.io/projected/18f99e6e-1055-4fef-b134-e49c03a0e06a-kube-api-access-6w8q2\") pod \"keystone-fa31-account-create-update-dbjwp\" (UID: \"18f99e6e-1055-4fef-b134-e49c03a0e06a\") " pod="openstack/keystone-fa31-account-create-update-dbjwp" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.522315 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35daefb6-843e-4791-8ef8-c9f850b7bf43-operator-scripts\") pod \"placement-db-create-ljfg7\" (UID: \"35daefb6-843e-4791-8ef8-c9f850b7bf43\") " pod="openstack/placement-db-create-ljfg7" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.523439 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18f99e6e-1055-4fef-b134-e49c03a0e06a-operator-scripts\") pod \"keystone-fa31-account-create-update-dbjwp\" (UID: \"18f99e6e-1055-4fef-b134-e49c03a0e06a\") " pod="openstack/keystone-fa31-account-create-update-dbjwp" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.547829 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w8q2\" (UniqueName: \"kubernetes.io/projected/18f99e6e-1055-4fef-b134-e49c03a0e06a-kube-api-access-6w8q2\") pod \"keystone-fa31-account-create-update-dbjwp\" (UID: \"18f99e6e-1055-4fef-b134-e49c03a0e06a\") " pod="openstack/keystone-fa31-account-create-update-dbjwp" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.624636 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35daefb6-843e-4791-8ef8-c9f850b7bf43-operator-scripts\") pod \"placement-db-create-ljfg7\" (UID: \"35daefb6-843e-4791-8ef8-c9f850b7bf43\") " pod="openstack/placement-db-create-ljfg7" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.624739 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25f2d927-0251-437d-a331-4a962f28f7bc-operator-scripts\") pod \"placement-73b1-account-create-update-8sddg\" (UID: \"25f2d927-0251-437d-a331-4a962f28f7bc\") " pod="openstack/placement-73b1-account-create-update-8sddg" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.624794 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r7lz\" (UniqueName: \"kubernetes.io/projected/35daefb6-843e-4791-8ef8-c9f850b7bf43-kube-api-access-6r7lz\") pod \"placement-db-create-ljfg7\" (UID: \"35daefb6-843e-4791-8ef8-c9f850b7bf43\") " pod="openstack/placement-db-create-ljfg7" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.624929 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhd46\" (UniqueName: \"kubernetes.io/projected/25f2d927-0251-437d-a331-4a962f28f7bc-kube-api-access-lhd46\") pod \"placement-73b1-account-create-update-8sddg\" (UID: \"25f2d927-0251-437d-a331-4a962f28f7bc\") " pod="openstack/placement-73b1-account-create-update-8sddg" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.625778 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35daefb6-843e-4791-8ef8-c9f850b7bf43-operator-scripts\") pod \"placement-db-create-ljfg7\" (UID: \"35daefb6-843e-4791-8ef8-c9f850b7bf43\") " pod="openstack/placement-db-create-ljfg7" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.644739 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r7lz\" (UniqueName: \"kubernetes.io/projected/35daefb6-843e-4791-8ef8-c9f850b7bf43-kube-api-access-6r7lz\") pod \"placement-db-create-ljfg7\" (UID: \"35daefb6-843e-4791-8ef8-c9f850b7bf43\") " pod="openstack/placement-db-create-ljfg7" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.726935 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhd46\" (UniqueName: \"kubernetes.io/projected/25f2d927-0251-437d-a331-4a962f28f7bc-kube-api-access-lhd46\") pod \"placement-73b1-account-create-update-8sddg\" (UID: \"25f2d927-0251-437d-a331-4a962f28f7bc\") " pod="openstack/placement-73b1-account-create-update-8sddg" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.727093 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25f2d927-0251-437d-a331-4a962f28f7bc-operator-scripts\") pod \"placement-73b1-account-create-update-8sddg\" (UID: \"25f2d927-0251-437d-a331-4a962f28f7bc\") " pod="openstack/placement-73b1-account-create-update-8sddg" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.727996 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25f2d927-0251-437d-a331-4a962f28f7bc-operator-scripts\") pod \"placement-73b1-account-create-update-8sddg\" (UID: \"25f2d927-0251-437d-a331-4a962f28f7bc\") " pod="openstack/placement-73b1-account-create-update-8sddg" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.748015 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhd46\" (UniqueName: \"kubernetes.io/projected/25f2d927-0251-437d-a331-4a962f28f7bc-kube-api-access-lhd46\") pod \"placement-73b1-account-create-update-8sddg\" (UID: \"25f2d927-0251-437d-a331-4a962f28f7bc\") " pod="openstack/placement-73b1-account-create-update-8sddg" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.841835 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fa31-account-create-update-dbjwp" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.892232 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ljfg7" Dec 06 07:15:25 crc kubenswrapper[4945]: I1206 07:15:25.894383 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-73b1-account-create-update-8sddg" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.088671 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-5zkfl"] Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.090066 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5zkfl" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.199536 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5zkfl"] Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.202036 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-f244-account-create-update-gvnlc"] Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.203531 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f244-account-create-update-gvnlc" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.207174 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.227190 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f244-account-create-update-gvnlc"] Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.237643 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/593549fe-1eb9-4e90-8a87-b2f620959e98-operator-scripts\") pod \"glance-db-create-5zkfl\" (UID: \"593549fe-1eb9-4e90-8a87-b2f620959e98\") " pod="openstack/glance-db-create-5zkfl" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.237747 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bgpk\" (UniqueName: \"kubernetes.io/projected/593549fe-1eb9-4e90-8a87-b2f620959e98-kube-api-access-6bgpk\") pod \"glance-db-create-5zkfl\" (UID: \"593549fe-1eb9-4e90-8a87-b2f620959e98\") " pod="openstack/glance-db-create-5zkfl" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.317592 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"250c85e2-e14d-4f93-822c-c9a88352eafe","Type":"ContainerStarted","Data":"5a233bae071c35e0cff40c7e5f127d5016758c87b7fe65dcf4b53dce1d196ba5"} Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.319158 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.332743 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-m5ptv" event={"ID":"5e2897e5-a177-4cb5-9790-4fa9bbb35104","Type":"ContainerStarted","Data":"6af3e0e431d8760ff4b89dcefc9f22ce56f5f677ccbf93f887d5053936a15a72"} Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.336648 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d71d0084-b4bc-4be5-971b-eced6e6d66ea","Type":"ContainerStarted","Data":"35755191ab63b793ba68c4e84cc61180235cdedf764a176765a67c7090121592"} Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.337485 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.340188 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/593549fe-1eb9-4e90-8a87-b2f620959e98-operator-scripts\") pod \"glance-db-create-5zkfl\" (UID: \"593549fe-1eb9-4e90-8a87-b2f620959e98\") " pod="openstack/glance-db-create-5zkfl" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.340243 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skss7\" (UniqueName: \"kubernetes.io/projected/69560226-9548-42f0-8c4a-cf33fed9b1f6-kube-api-access-skss7\") pod \"glance-f244-account-create-update-gvnlc\" (UID: \"69560226-9548-42f0-8c4a-cf33fed9b1f6\") " pod="openstack/glance-f244-account-create-update-gvnlc" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.340301 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bgpk\" (UniqueName: \"kubernetes.io/projected/593549fe-1eb9-4e90-8a87-b2f620959e98-kube-api-access-6bgpk\") pod \"glance-db-create-5zkfl\" (UID: \"593549fe-1eb9-4e90-8a87-b2f620959e98\") " pod="openstack/glance-db-create-5zkfl" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.340354 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69560226-9548-42f0-8c4a-cf33fed9b1f6-operator-scripts\") pod \"glance-f244-account-create-update-gvnlc\" (UID: \"69560226-9548-42f0-8c4a-cf33fed9b1f6\") " pod="openstack/glance-f244-account-create-update-gvnlc" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.341839 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/593549fe-1eb9-4e90-8a87-b2f620959e98-operator-scripts\") pod \"glance-db-create-5zkfl\" (UID: \"593549fe-1eb9-4e90-8a87-b2f620959e98\") " pod="openstack/glance-db-create-5zkfl" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.370470 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.112922887 podStartE2EDuration="1m16.370445204s" podCreationTimestamp="2025-12-06 07:14:10 +0000 UTC" firstStartedPulling="2025-12-06 07:14:12.53638464 +0000 UTC m=+1265.991245684" lastFinishedPulling="2025-12-06 07:14:50.793906957 +0000 UTC m=+1304.248768001" observedRunningTime="2025-12-06 07:15:26.363810595 +0000 UTC m=+1339.818671659" watchObservedRunningTime="2025-12-06 07:15:26.370445204 +0000 UTC m=+1339.825306248" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.396225 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bgpk\" (UniqueName: \"kubernetes.io/projected/593549fe-1eb9-4e90-8a87-b2f620959e98-kube-api-access-6bgpk\") pod \"glance-db-create-5zkfl\" (UID: \"593549fe-1eb9-4e90-8a87-b2f620959e98\") " pod="openstack/glance-db-create-5zkfl" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.415758 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-m5ptv" podStartSLOduration=2.415732509 podStartE2EDuration="2.415732509s" podCreationTimestamp="2025-12-06 07:15:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:26.396478138 +0000 UTC m=+1339.851339192" watchObservedRunningTime="2025-12-06 07:15:26.415732509 +0000 UTC m=+1339.870593553" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.416263 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5zkfl" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.431524 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.790801665000004 podStartE2EDuration="1m17.431508056s" podCreationTimestamp="2025-12-06 07:14:09 +0000 UTC" firstStartedPulling="2025-12-06 07:14:12.109144344 +0000 UTC m=+1265.564005388" lastFinishedPulling="2025-12-06 07:14:50.749850735 +0000 UTC m=+1304.204711779" observedRunningTime="2025-12-06 07:15:26.430883759 +0000 UTC m=+1339.885744803" watchObservedRunningTime="2025-12-06 07:15:26.431508056 +0000 UTC m=+1339.886369100" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.442663 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skss7\" (UniqueName: \"kubernetes.io/projected/69560226-9548-42f0-8c4a-cf33fed9b1f6-kube-api-access-skss7\") pod \"glance-f244-account-create-update-gvnlc\" (UID: \"69560226-9548-42f0-8c4a-cf33fed9b1f6\") " pod="openstack/glance-f244-account-create-update-gvnlc" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.442837 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69560226-9548-42f0-8c4a-cf33fed9b1f6-operator-scripts\") pod \"glance-f244-account-create-update-gvnlc\" (UID: \"69560226-9548-42f0-8c4a-cf33fed9b1f6\") " pod="openstack/glance-f244-account-create-update-gvnlc" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.446031 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69560226-9548-42f0-8c4a-cf33fed9b1f6-operator-scripts\") pod \"glance-f244-account-create-update-gvnlc\" (UID: \"69560226-9548-42f0-8c4a-cf33fed9b1f6\") " pod="openstack/glance-f244-account-create-update-gvnlc" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.486434 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skss7\" (UniqueName: \"kubernetes.io/projected/69560226-9548-42f0-8c4a-cf33fed9b1f6-kube-api-access-skss7\") pod \"glance-f244-account-create-update-gvnlc\" (UID: \"69560226-9548-42f0-8c4a-cf33fed9b1f6\") " pod="openstack/glance-f244-account-create-update-gvnlc" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.582919 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f244-account-create-update-gvnlc" Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.849145 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fa31-account-create-update-dbjwp"] Dec 06 07:15:26 crc kubenswrapper[4945]: W1206 07:15:26.941604 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25f2d927_0251_437d_a331_4a962f28f7bc.slice/crio-b6d6bff40fa1a2131049f09148de16f5bc9310122aaf33a4de1680845fe73c32 WatchSource:0}: Error finding container b6d6bff40fa1a2131049f09148de16f5bc9310122aaf33a4de1680845fe73c32: Status 404 returned error can't find the container with id b6d6bff40fa1a2131049f09148de16f5bc9310122aaf33a4de1680845fe73c32 Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.942786 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-73b1-account-create-update-8sddg"] Dec 06 07:15:26 crc kubenswrapper[4945]: I1206 07:15:26.981131 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f244-account-create-update-gvnlc"] Dec 06 07:15:27 crc kubenswrapper[4945]: I1206 07:15:27.012488 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ljfg7"] Dec 06 07:15:27 crc kubenswrapper[4945]: I1206 07:15:27.083717 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5zkfl"] Dec 06 07:15:27 crc kubenswrapper[4945]: W1206 07:15:27.094950 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod593549fe_1eb9_4e90_8a87_b2f620959e98.slice/crio-1a451f43e6767a65d285ab276b50e4415bcfc1f335c205bddfaadb5c3d708285 WatchSource:0}: Error finding container 1a451f43e6767a65d285ab276b50e4415bcfc1f335c205bddfaadb5c3d708285: Status 404 returned error can't find the container with id 1a451f43e6767a65d285ab276b50e4415bcfc1f335c205bddfaadb5c3d708285 Dec 06 07:15:27 crc kubenswrapper[4945]: I1206 07:15:27.362512 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f244-account-create-update-gvnlc" event={"ID":"69560226-9548-42f0-8c4a-cf33fed9b1f6","Type":"ContainerStarted","Data":"f6d58864bfcd1a1dad883be27edcbb9374d3cd74f3d8fe1f1a1d8eac9207937c"} Dec 06 07:15:27 crc kubenswrapper[4945]: I1206 07:15:27.369870 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ljfg7" event={"ID":"35daefb6-843e-4791-8ef8-c9f850b7bf43","Type":"ContainerStarted","Data":"e90ec5621afdb6f34f87f45babab4e3ca4356cdb82b6ae2831fe959719744039"} Dec 06 07:15:27 crc kubenswrapper[4945]: I1206 07:15:27.371178 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-73b1-account-create-update-8sddg" event={"ID":"25f2d927-0251-437d-a331-4a962f28f7bc","Type":"ContainerStarted","Data":"b6d6bff40fa1a2131049f09148de16f5bc9310122aaf33a4de1680845fe73c32"} Dec 06 07:15:27 crc kubenswrapper[4945]: I1206 07:15:27.372371 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fa31-account-create-update-dbjwp" event={"ID":"18f99e6e-1055-4fef-b134-e49c03a0e06a","Type":"ContainerStarted","Data":"9ce3836033d9027cdd9d8a8c7eaeddfe859b7b9a03ec9c4c98818c8f607f7bad"} Dec 06 07:15:27 crc kubenswrapper[4945]: I1206 07:15:27.375106 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5zkfl" event={"ID":"593549fe-1eb9-4e90-8a87-b2f620959e98","Type":"ContainerStarted","Data":"1a451f43e6767a65d285ab276b50e4415bcfc1f335c205bddfaadb5c3d708285"} Dec 06 07:15:28 crc kubenswrapper[4945]: I1206 07:15:28.398913 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f244-account-create-update-gvnlc" event={"ID":"69560226-9548-42f0-8c4a-cf33fed9b1f6","Type":"ContainerStarted","Data":"328607c02897e7568c645c5ee8ef89528ebfeb9ca65456bea19a7c18881606e6"} Dec 06 07:15:28 crc kubenswrapper[4945]: I1206 07:15:28.405967 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ljfg7" event={"ID":"35daefb6-843e-4791-8ef8-c9f850b7bf43","Type":"ContainerStarted","Data":"453ba68949233d03da1550a4fe35605ece36d1e51233e8fd776fbc8f82477d6c"} Dec 06 07:15:28 crc kubenswrapper[4945]: I1206 07:15:28.418810 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-73b1-account-create-update-8sddg" event={"ID":"25f2d927-0251-437d-a331-4a962f28f7bc","Type":"ContainerStarted","Data":"1d208203ca7f66b7e885206b007a585ace9805113b2f6c3f8842bd6232730d75"} Dec 06 07:15:28 crc kubenswrapper[4945]: I1206 07:15:28.429849 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fa31-account-create-update-dbjwp" event={"ID":"18f99e6e-1055-4fef-b134-e49c03a0e06a","Type":"ContainerStarted","Data":"3bd788957d65014f6df7ecdbf42af33341027a132b2359cc1089be7edcca1dc3"} Dec 06 07:15:28 crc kubenswrapper[4945]: I1206 07:15:28.430942 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-f244-account-create-update-gvnlc" podStartSLOduration=2.430918383 podStartE2EDuration="2.430918383s" podCreationTimestamp="2025-12-06 07:15:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:28.425210519 +0000 UTC m=+1341.880071583" watchObservedRunningTime="2025-12-06 07:15:28.430918383 +0000 UTC m=+1341.885779427" Dec 06 07:15:28 crc kubenswrapper[4945]: I1206 07:15:28.444918 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5zkfl" event={"ID":"593549fe-1eb9-4e90-8a87-b2f620959e98","Type":"ContainerStarted","Data":"40ab40310e6772f4b9f2541373cf860b3422560fe3fcc29e1733da457cd5d585"} Dec 06 07:15:28 crc kubenswrapper[4945]: I1206 07:15:28.481538 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-73b1-account-create-update-8sddg" podStartSLOduration=3.481518932 podStartE2EDuration="3.481518932s" podCreationTimestamp="2025-12-06 07:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:28.451414647 +0000 UTC m=+1341.906275701" watchObservedRunningTime="2025-12-06 07:15:28.481518932 +0000 UTC m=+1341.936379986" Dec 06 07:15:28 crc kubenswrapper[4945]: I1206 07:15:28.485232 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-ljfg7" podStartSLOduration=3.485222672 podStartE2EDuration="3.485222672s" podCreationTimestamp="2025-12-06 07:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:28.477817411 +0000 UTC m=+1341.932678455" watchObservedRunningTime="2025-12-06 07:15:28.485222672 +0000 UTC m=+1341.940083716" Dec 06 07:15:28 crc kubenswrapper[4945]: I1206 07:15:28.501420 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-5zkfl" podStartSLOduration=2.5013673279999997 podStartE2EDuration="2.501367328s" podCreationTimestamp="2025-12-06 07:15:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:28.496056075 +0000 UTC m=+1341.950917119" watchObservedRunningTime="2025-12-06 07:15:28.501367328 +0000 UTC m=+1341.956228372" Dec 06 07:15:28 crc kubenswrapper[4945]: I1206 07:15:28.530612 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-fa31-account-create-update-dbjwp" podStartSLOduration=3.530593239 podStartE2EDuration="3.530593239s" podCreationTimestamp="2025-12-06 07:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:28.528857512 +0000 UTC m=+1341.983718556" watchObservedRunningTime="2025-12-06 07:15:28.530593239 +0000 UTC m=+1341.985454283" Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.456613 4945 generic.go:334] "Generic (PLEG): container finished" podID="18f99e6e-1055-4fef-b134-e49c03a0e06a" containerID="3bd788957d65014f6df7ecdbf42af33341027a132b2359cc1089be7edcca1dc3" exitCode=0 Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.456827 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fa31-account-create-update-dbjwp" event={"ID":"18f99e6e-1055-4fef-b134-e49c03a0e06a","Type":"ContainerDied","Data":"3bd788957d65014f6df7ecdbf42af33341027a132b2359cc1089be7edcca1dc3"} Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.459104 4945 generic.go:334] "Generic (PLEG): container finished" podID="593549fe-1eb9-4e90-8a87-b2f620959e98" containerID="40ab40310e6772f4b9f2541373cf860b3422560fe3fcc29e1733da457cd5d585" exitCode=0 Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.459172 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5zkfl" event={"ID":"593549fe-1eb9-4e90-8a87-b2f620959e98","Type":"ContainerDied","Data":"40ab40310e6772f4b9f2541373cf860b3422560fe3fcc29e1733da457cd5d585"} Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.461200 4945 generic.go:334] "Generic (PLEG): container finished" podID="69560226-9548-42f0-8c4a-cf33fed9b1f6" containerID="328607c02897e7568c645c5ee8ef89528ebfeb9ca65456bea19a7c18881606e6" exitCode=0 Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.461253 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f244-account-create-update-gvnlc" event={"ID":"69560226-9548-42f0-8c4a-cf33fed9b1f6","Type":"ContainerDied","Data":"328607c02897e7568c645c5ee8ef89528ebfeb9ca65456bea19a7c18881606e6"} Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.463033 4945 generic.go:334] "Generic (PLEG): container finished" podID="35daefb6-843e-4791-8ef8-c9f850b7bf43" containerID="453ba68949233d03da1550a4fe35605ece36d1e51233e8fd776fbc8f82477d6c" exitCode=0 Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.463080 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ljfg7" event={"ID":"35daefb6-843e-4791-8ef8-c9f850b7bf43","Type":"ContainerDied","Data":"453ba68949233d03da1550a4fe35605ece36d1e51233e8fd776fbc8f82477d6c"} Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.464652 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e2897e5-a177-4cb5-9790-4fa9bbb35104" containerID="6af3e0e431d8760ff4b89dcefc9f22ce56f5f677ccbf93f887d5053936a15a72" exitCode=0 Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.464702 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-m5ptv" event={"ID":"5e2897e5-a177-4cb5-9790-4fa9bbb35104","Type":"ContainerDied","Data":"6af3e0e431d8760ff4b89dcefc9f22ce56f5f677ccbf93f887d5053936a15a72"} Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.466762 4945 generic.go:334] "Generic (PLEG): container finished" podID="25f2d927-0251-437d-a331-4a962f28f7bc" containerID="1d208203ca7f66b7e885206b007a585ace9805113b2f6c3f8842bd6232730d75" exitCode=0 Dec 06 07:15:29 crc kubenswrapper[4945]: I1206 07:15:29.466791 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-73b1-account-create-update-8sddg" event={"ID":"25f2d927-0251-437d-a331-4a962f28f7bc","Type":"ContainerDied","Data":"1d208203ca7f66b7e885206b007a585ace9805113b2f6c3f8842bd6232730d75"} Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.132319 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fa31-account-create-update-dbjwp" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.206965 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xc2dd" podUID="3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" containerName="ovn-controller" probeResult="failure" output=< Dec 06 07:15:31 crc kubenswrapper[4945]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 07:15:31 crc kubenswrapper[4945]: > Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.226168 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.244960 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w8q2\" (UniqueName: \"kubernetes.io/projected/18f99e6e-1055-4fef-b134-e49c03a0e06a-kube-api-access-6w8q2\") pod \"18f99e6e-1055-4fef-b134-e49c03a0e06a\" (UID: \"18f99e6e-1055-4fef-b134-e49c03a0e06a\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.245161 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18f99e6e-1055-4fef-b134-e49c03a0e06a-operator-scripts\") pod \"18f99e6e-1055-4fef-b134-e49c03a0e06a\" (UID: \"18f99e6e-1055-4fef-b134-e49c03a0e06a\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.246232 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18f99e6e-1055-4fef-b134-e49c03a0e06a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "18f99e6e-1055-4fef-b134-e49c03a0e06a" (UID: "18f99e6e-1055-4fef-b134-e49c03a0e06a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.247701 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.279111 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18f99e6e-1055-4fef-b134-e49c03a0e06a-kube-api-access-6w8q2" (OuterVolumeSpecName: "kube-api-access-6w8q2") pod "18f99e6e-1055-4fef-b134-e49c03a0e06a" (UID: "18f99e6e-1055-4fef-b134-e49c03a0e06a"). InnerVolumeSpecName "kube-api-access-6w8q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.347888 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w8q2\" (UniqueName: \"kubernetes.io/projected/18f99e6e-1055-4fef-b134-e49c03a0e06a-kube-api-access-6w8q2\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.347939 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18f99e6e-1055-4fef-b134-e49c03a0e06a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.355776 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ljfg7" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.438725 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m5ptv" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.448951 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35daefb6-843e-4791-8ef8-c9f850b7bf43-operator-scripts\") pod \"35daefb6-843e-4791-8ef8-c9f850b7bf43\" (UID: \"35daefb6-843e-4791-8ef8-c9f850b7bf43\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.448992 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r7lz\" (UniqueName: \"kubernetes.io/projected/35daefb6-843e-4791-8ef8-c9f850b7bf43-kube-api-access-6r7lz\") pod \"35daefb6-843e-4791-8ef8-c9f850b7bf43\" (UID: \"35daefb6-843e-4791-8ef8-c9f850b7bf43\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.450095 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35daefb6-843e-4791-8ef8-c9f850b7bf43-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "35daefb6-843e-4791-8ef8-c9f850b7bf43" (UID: "35daefb6-843e-4791-8ef8-c9f850b7bf43"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.450487 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35daefb6-843e-4791-8ef8-c9f850b7bf43-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.452931 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f244-account-create-update-gvnlc" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.454335 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35daefb6-843e-4791-8ef8-c9f850b7bf43-kube-api-access-6r7lz" (OuterVolumeSpecName: "kube-api-access-6r7lz") pod "35daefb6-843e-4791-8ef8-c9f850b7bf43" (UID: "35daefb6-843e-4791-8ef8-c9f850b7bf43"). InnerVolumeSpecName "kube-api-access-6r7lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.462801 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-73b1-account-create-update-8sddg" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.479687 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5zkfl" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.494636 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f244-account-create-update-gvnlc" event={"ID":"69560226-9548-42f0-8c4a-cf33fed9b1f6","Type":"ContainerDied","Data":"f6d58864bfcd1a1dad883be27edcbb9374d3cd74f3d8fe1f1a1d8eac9207937c"} Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.494695 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6d58864bfcd1a1dad883be27edcbb9374d3cd74f3d8fe1f1a1d8eac9207937c" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.494664 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f244-account-create-update-gvnlc" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.496355 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ljfg7" event={"ID":"35daefb6-843e-4791-8ef8-c9f850b7bf43","Type":"ContainerDied","Data":"e90ec5621afdb6f34f87f45babab4e3ca4356cdb82b6ae2831fe959719744039"} Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.496405 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e90ec5621afdb6f34f87f45babab4e3ca4356cdb82b6ae2831fe959719744039" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.496474 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ljfg7" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.503726 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-m5ptv" event={"ID":"5e2897e5-a177-4cb5-9790-4fa9bbb35104","Type":"ContainerDied","Data":"6182a1bbfd1ca2dcd498015cec1d5509c804d28281c1e46ec96a2d57a160b99a"} Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.503989 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m5ptv" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.504086 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6182a1bbfd1ca2dcd498015cec1d5509c804d28281c1e46ec96a2d57a160b99a" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.530738 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-73b1-account-create-update-8sddg" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.531143 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-73b1-account-create-update-8sddg" event={"ID":"25f2d927-0251-437d-a331-4a962f28f7bc","Type":"ContainerDied","Data":"b6d6bff40fa1a2131049f09148de16f5bc9310122aaf33a4de1680845fe73c32"} Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.531174 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6d6bff40fa1a2131049f09148de16f5bc9310122aaf33a4de1680845fe73c32" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.548664 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fa31-account-create-update-dbjwp" event={"ID":"18f99e6e-1055-4fef-b134-e49c03a0e06a","Type":"ContainerDied","Data":"9ce3836033d9027cdd9d8a8c7eaeddfe859b7b9a03ec9c4c98818c8f607f7bad"} Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.548725 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ce3836033d9027cdd9d8a8c7eaeddfe859b7b9a03ec9c4c98818c8f607f7bad" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.548689 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fa31-account-create-update-dbjwp" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.554338 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xc2dd-config-mndst"] Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.555160 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25f2d927-0251-437d-a331-4a962f28f7bc-operator-scripts\") pod \"25f2d927-0251-437d-a331-4a962f28f7bc\" (UID: \"25f2d927-0251-437d-a331-4a962f28f7bc\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.559171 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/593549fe-1eb9-4e90-8a87-b2f620959e98-operator-scripts\") pod \"593549fe-1eb9-4e90-8a87-b2f620959e98\" (UID: \"593549fe-1eb9-4e90-8a87-b2f620959e98\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.559395 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e2897e5-a177-4cb5-9790-4fa9bbb35104-operator-scripts\") pod \"5e2897e5-a177-4cb5-9790-4fa9bbb35104\" (UID: \"5e2897e5-a177-4cb5-9790-4fa9bbb35104\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.559504 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skss7\" (UniqueName: \"kubernetes.io/projected/69560226-9548-42f0-8c4a-cf33fed9b1f6-kube-api-access-skss7\") pod \"69560226-9548-42f0-8c4a-cf33fed9b1f6\" (UID: \"69560226-9548-42f0-8c4a-cf33fed9b1f6\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.559584 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhd46\" (UniqueName: \"kubernetes.io/projected/25f2d927-0251-437d-a331-4a962f28f7bc-kube-api-access-lhd46\") pod \"25f2d927-0251-437d-a331-4a962f28f7bc\" (UID: \"25f2d927-0251-437d-a331-4a962f28f7bc\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.560129 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69560226-9548-42f0-8c4a-cf33fed9b1f6-operator-scripts\") pod \"69560226-9548-42f0-8c4a-cf33fed9b1f6\" (UID: \"69560226-9548-42f0-8c4a-cf33fed9b1f6\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.560204 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bgpk\" (UniqueName: \"kubernetes.io/projected/593549fe-1eb9-4e90-8a87-b2f620959e98-kube-api-access-6bgpk\") pod \"593549fe-1eb9-4e90-8a87-b2f620959e98\" (UID: \"593549fe-1eb9-4e90-8a87-b2f620959e98\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.560244 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqzrm\" (UniqueName: \"kubernetes.io/projected/5e2897e5-a177-4cb5-9790-4fa9bbb35104-kube-api-access-bqzrm\") pod \"5e2897e5-a177-4cb5-9790-4fa9bbb35104\" (UID: \"5e2897e5-a177-4cb5-9790-4fa9bbb35104\") " Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.561543 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25f2d927-0251-437d-a331-4a962f28f7bc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "25f2d927-0251-437d-a331-4a962f28f7bc" (UID: "25f2d927-0251-437d-a331-4a962f28f7bc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.573920 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/593549fe-1eb9-4e90-8a87-b2f620959e98-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "593549fe-1eb9-4e90-8a87-b2f620959e98" (UID: "593549fe-1eb9-4e90-8a87-b2f620959e98"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.576531 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e2897e5-a177-4cb5-9790-4fa9bbb35104-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5e2897e5-a177-4cb5-9790-4fa9bbb35104" (UID: "5e2897e5-a177-4cb5-9790-4fa9bbb35104"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.578475 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69560226-9548-42f0-8c4a-cf33fed9b1f6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "69560226-9548-42f0-8c4a-cf33fed9b1f6" (UID: "69560226-9548-42f0-8c4a-cf33fed9b1f6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.582359 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25f2d927-0251-437d-a331-4a962f28f7bc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.582395 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/593549fe-1eb9-4e90-8a87-b2f620959e98-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.582419 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r7lz\" (UniqueName: \"kubernetes.io/projected/35daefb6-843e-4791-8ef8-c9f850b7bf43-kube-api-access-6r7lz\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.582433 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e2897e5-a177-4cb5-9790-4fa9bbb35104-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.582443 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69560226-9548-42f0-8c4a-cf33fed9b1f6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.582583 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25f2d927-0251-437d-a331-4a962f28f7bc-kube-api-access-lhd46" (OuterVolumeSpecName: "kube-api-access-lhd46") pod "25f2d927-0251-437d-a331-4a962f28f7bc" (UID: "25f2d927-0251-437d-a331-4a962f28f7bc"). InnerVolumeSpecName "kube-api-access-lhd46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.582838 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5zkfl" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.583922 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/593549fe-1eb9-4e90-8a87-b2f620959e98-kube-api-access-6bgpk" (OuterVolumeSpecName: "kube-api-access-6bgpk") pod "593549fe-1eb9-4e90-8a87-b2f620959e98" (UID: "593549fe-1eb9-4e90-8a87-b2f620959e98"). InnerVolumeSpecName "kube-api-access-6bgpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: E1206 07:15:31.584327 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2897e5-a177-4cb5-9790-4fa9bbb35104" containerName="mariadb-database-create" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.584349 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2897e5-a177-4cb5-9790-4fa9bbb35104" containerName="mariadb-database-create" Dec 06 07:15:31 crc kubenswrapper[4945]: E1206 07:15:31.584391 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f99e6e-1055-4fef-b134-e49c03a0e06a" containerName="mariadb-account-create-update" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.584400 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f99e6e-1055-4fef-b134-e49c03a0e06a" containerName="mariadb-account-create-update" Dec 06 07:15:31 crc kubenswrapper[4945]: E1206 07:15:31.584410 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="593549fe-1eb9-4e90-8a87-b2f620959e98" containerName="mariadb-database-create" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.584416 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="593549fe-1eb9-4e90-8a87-b2f620959e98" containerName="mariadb-database-create" Dec 06 07:15:31 crc kubenswrapper[4945]: E1206 07:15:31.584439 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69560226-9548-42f0-8c4a-cf33fed9b1f6" containerName="mariadb-account-create-update" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.584445 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="69560226-9548-42f0-8c4a-cf33fed9b1f6" containerName="mariadb-account-create-update" Dec 06 07:15:31 crc kubenswrapper[4945]: E1206 07:15:31.584460 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35daefb6-843e-4791-8ef8-c9f850b7bf43" containerName="mariadb-database-create" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.584466 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="35daefb6-843e-4791-8ef8-c9f850b7bf43" containerName="mariadb-database-create" Dec 06 07:15:31 crc kubenswrapper[4945]: E1206 07:15:31.584484 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f2d927-0251-437d-a331-4a962f28f7bc" containerName="mariadb-account-create-update" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.584491 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f2d927-0251-437d-a331-4a962f28f7bc" containerName="mariadb-account-create-update" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.585873 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2897e5-a177-4cb5-9790-4fa9bbb35104" containerName="mariadb-database-create" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.585898 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="593549fe-1eb9-4e90-8a87-b2f620959e98" containerName="mariadb-database-create" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.585937 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="35daefb6-843e-4791-8ef8-c9f850b7bf43" containerName="mariadb-database-create" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.586028 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="69560226-9548-42f0-8c4a-cf33fed9b1f6" containerName="mariadb-account-create-update" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.586047 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="18f99e6e-1055-4fef-b134-e49c03a0e06a" containerName="mariadb-account-create-update" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.586092 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="25f2d927-0251-437d-a331-4a962f28f7bc" containerName="mariadb-account-create-update" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.590182 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5zkfl" event={"ID":"593549fe-1eb9-4e90-8a87-b2f620959e98","Type":"ContainerDied","Data":"1a451f43e6767a65d285ab276b50e4415bcfc1f335c205bddfaadb5c3d708285"} Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.590216 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a451f43e6767a65d285ab276b50e4415bcfc1f335c205bddfaadb5c3d708285" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.590332 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.591817 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2897e5-a177-4cb5-9790-4fa9bbb35104-kube-api-access-bqzrm" (OuterVolumeSpecName: "kube-api-access-bqzrm") pod "5e2897e5-a177-4cb5-9790-4fa9bbb35104" (UID: "5e2897e5-a177-4cb5-9790-4fa9bbb35104"). InnerVolumeSpecName "kube-api-access-bqzrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.593766 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.600491 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69560226-9548-42f0-8c4a-cf33fed9b1f6-kube-api-access-skss7" (OuterVolumeSpecName: "kube-api-access-skss7") pod "69560226-9548-42f0-8c4a-cf33fed9b1f6" (UID: "69560226-9548-42f0-8c4a-cf33fed9b1f6"). InnerVolumeSpecName "kube-api-access-skss7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.650314 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xc2dd-config-mndst"] Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.684769 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.684818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-scripts\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.684855 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run-ovn\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.685170 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-log-ovn\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.685233 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r69s\" (UniqueName: \"kubernetes.io/projected/f67adcdc-2163-48d0-b9d4-013d03f48b5b-kube-api-access-2r69s\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.685264 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-additional-scripts\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.685653 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skss7\" (UniqueName: \"kubernetes.io/projected/69560226-9548-42f0-8c4a-cf33fed9b1f6-kube-api-access-skss7\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.685697 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhd46\" (UniqueName: \"kubernetes.io/projected/25f2d927-0251-437d-a331-4a962f28f7bc-kube-api-access-lhd46\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.685711 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bgpk\" (UniqueName: \"kubernetes.io/projected/593549fe-1eb9-4e90-8a87-b2f620959e98-kube-api-access-6bgpk\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.685722 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqzrm\" (UniqueName: \"kubernetes.io/projected/5e2897e5-a177-4cb5-9790-4fa9bbb35104-kube-api-access-bqzrm\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.787198 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r69s\" (UniqueName: \"kubernetes.io/projected/f67adcdc-2163-48d0-b9d4-013d03f48b5b-kube-api-access-2r69s\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.787248 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-additional-scripts\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.787424 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.787465 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-scripts\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.787485 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run-ovn\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.787530 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-log-ovn\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.787822 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-log-ovn\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.787842 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.788028 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run-ovn\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.789106 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-additional-scripts\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.790194 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-scripts\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.803512 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r69s\" (UniqueName: \"kubernetes.io/projected/f67adcdc-2163-48d0-b9d4-013d03f48b5b-kube-api-access-2r69s\") pod \"ovn-controller-xc2dd-config-mndst\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:31 crc kubenswrapper[4945]: I1206 07:15:31.932833 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:32 crc kubenswrapper[4945]: I1206 07:15:32.771877 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xc2dd-config-mndst"] Dec 06 07:15:33 crc kubenswrapper[4945]: I1206 07:15:33.604891 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xc2dd-config-mndst" event={"ID":"f67adcdc-2163-48d0-b9d4-013d03f48b5b","Type":"ContainerStarted","Data":"27630bdcc77303b93d309a9fbb7dc77578d594516e2a965d251b84f4c0aa821f"} Dec 06 07:15:33 crc kubenswrapper[4945]: I1206 07:15:33.605204 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xc2dd-config-mndst" event={"ID":"f67adcdc-2163-48d0-b9d4-013d03f48b5b","Type":"ContainerStarted","Data":"ca2fa9000eea4c21e9583b5786e56ca95615fff5b580a46b7bd10170816f387a"} Dec 06 07:15:33 crc kubenswrapper[4945]: I1206 07:15:33.645511 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xc2dd-config-mndst" podStartSLOduration=2.645480808 podStartE2EDuration="2.645480808s" podCreationTimestamp="2025-12-06 07:15:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:33.626888185 +0000 UTC m=+1347.081749249" watchObservedRunningTime="2025-12-06 07:15:33.645480808 +0000 UTC m=+1347.100341872" Dec 06 07:15:34 crc kubenswrapper[4945]: I1206 07:15:34.615062 4945 generic.go:334] "Generic (PLEG): container finished" podID="f67adcdc-2163-48d0-b9d4-013d03f48b5b" containerID="27630bdcc77303b93d309a9fbb7dc77578d594516e2a965d251b84f4c0aa821f" exitCode=0 Dec 06 07:15:34 crc kubenswrapper[4945]: I1206 07:15:34.615167 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xc2dd-config-mndst" event={"ID":"f67adcdc-2163-48d0-b9d4-013d03f48b5b","Type":"ContainerDied","Data":"27630bdcc77303b93d309a9fbb7dc77578d594516e2a965d251b84f4c0aa821f"} Dec 06 07:15:34 crc kubenswrapper[4945]: I1206 07:15:34.618728 4945 generic.go:334] "Generic (PLEG): container finished" podID="3603082b-401e-484c-88a3-90c8f67f5824" containerID="3624afb9c68cb040abd682f4ced6ee22803012bbc4d7d8b26dd3e7ce553edb4c" exitCode=0 Dec 06 07:15:34 crc kubenswrapper[4945]: I1206 07:15:34.618773 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hjh59" event={"ID":"3603082b-401e-484c-88a3-90c8f67f5824","Type":"ContainerDied","Data":"3624afb9c68cb040abd682f4ced6ee22803012bbc4d7d8b26dd3e7ce553edb4c"} Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.166501 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-xc2dd" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.177685 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.263265 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-combined-ca-bundle\") pod \"3603082b-401e-484c-88a3-90c8f67f5824\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.263388 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-swiftconf\") pod \"3603082b-401e-484c-88a3-90c8f67f5824\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.263495 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3603082b-401e-484c-88a3-90c8f67f5824-etc-swift\") pod \"3603082b-401e-484c-88a3-90c8f67f5824\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.266334 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-scripts\") pod \"3603082b-401e-484c-88a3-90c8f67f5824\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.266498 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-dispersionconf\") pod \"3603082b-401e-484c-88a3-90c8f67f5824\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.266570 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktmfs\" (UniqueName: \"kubernetes.io/projected/3603082b-401e-484c-88a3-90c8f67f5824-kube-api-access-ktmfs\") pod \"3603082b-401e-484c-88a3-90c8f67f5824\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.266619 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-ring-data-devices\") pod \"3603082b-401e-484c-88a3-90c8f67f5824\" (UID: \"3603082b-401e-484c-88a3-90c8f67f5824\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.266724 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3603082b-401e-484c-88a3-90c8f67f5824-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3603082b-401e-484c-88a3-90c8f67f5824" (UID: "3603082b-401e-484c-88a3-90c8f67f5824"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.267715 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3603082b-401e-484c-88a3-90c8f67f5824" (UID: "3603082b-401e-484c-88a3-90c8f67f5824"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.270924 4945 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.271200 4945 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3603082b-401e-484c-88a3-90c8f67f5824-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.278913 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3603082b-401e-484c-88a3-90c8f67f5824-kube-api-access-ktmfs" (OuterVolumeSpecName: "kube-api-access-ktmfs") pod "3603082b-401e-484c-88a3-90c8f67f5824" (UID: "3603082b-401e-484c-88a3-90c8f67f5824"). InnerVolumeSpecName "kube-api-access-ktmfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.289936 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3603082b-401e-484c-88a3-90c8f67f5824" (UID: "3603082b-401e-484c-88a3-90c8f67f5824"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.308079 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3603082b-401e-484c-88a3-90c8f67f5824" (UID: "3603082b-401e-484c-88a3-90c8f67f5824"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.318903 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3603082b-401e-484c-88a3-90c8f67f5824" (UID: "3603082b-401e-484c-88a3-90c8f67f5824"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.326393 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-scripts" (OuterVolumeSpecName: "scripts") pod "3603082b-401e-484c-88a3-90c8f67f5824" (UID: "3603082b-401e-484c-88a3-90c8f67f5824"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.348695 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.373477 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktmfs\" (UniqueName: \"kubernetes.io/projected/3603082b-401e-484c-88a3-90c8f67f5824-kube-api-access-ktmfs\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.373514 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.373525 4945 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.373534 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3603082b-401e-484c-88a3-90c8f67f5824-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.373546 4945 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3603082b-401e-484c-88a3-90c8f67f5824-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.474272 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-log-ovn\") pod \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.474400 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run\") pod \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.474402 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f67adcdc-2163-48d0-b9d4-013d03f48b5b" (UID: "f67adcdc-2163-48d0-b9d4-013d03f48b5b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.474499 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run" (OuterVolumeSpecName: "var-run") pod "f67adcdc-2163-48d0-b9d4-013d03f48b5b" (UID: "f67adcdc-2163-48d0-b9d4-013d03f48b5b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.474560 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-scripts\") pod \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.475952 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-scripts" (OuterVolumeSpecName: "scripts") pod "f67adcdc-2163-48d0-b9d4-013d03f48b5b" (UID: "f67adcdc-2163-48d0-b9d4-013d03f48b5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.476012 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r69s\" (UniqueName: \"kubernetes.io/projected/f67adcdc-2163-48d0-b9d4-013d03f48b5b-kube-api-access-2r69s\") pod \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.476144 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run-ovn\") pod \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.476302 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-additional-scripts\") pod \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\" (UID: \"f67adcdc-2163-48d0-b9d4-013d03f48b5b\") " Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.476845 4945 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.476861 4945 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.476874 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.477207 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f67adcdc-2163-48d0-b9d4-013d03f48b5b" (UID: "f67adcdc-2163-48d0-b9d4-013d03f48b5b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.477594 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f67adcdc-2163-48d0-b9d4-013d03f48b5b" (UID: "f67adcdc-2163-48d0-b9d4-013d03f48b5b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.480530 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f67adcdc-2163-48d0-b9d4-013d03f48b5b-kube-api-access-2r69s" (OuterVolumeSpecName: "kube-api-access-2r69s") pod "f67adcdc-2163-48d0-b9d4-013d03f48b5b" (UID: "f67adcdc-2163-48d0-b9d4-013d03f48b5b"). InnerVolumeSpecName "kube-api-access-2r69s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.509085 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-g5p9c"] Dec 06 07:15:36 crc kubenswrapper[4945]: E1206 07:15:36.509559 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67adcdc-2163-48d0-b9d4-013d03f48b5b" containerName="ovn-config" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.509586 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67adcdc-2163-48d0-b9d4-013d03f48b5b" containerName="ovn-config" Dec 06 07:15:36 crc kubenswrapper[4945]: E1206 07:15:36.509624 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3603082b-401e-484c-88a3-90c8f67f5824" containerName="swift-ring-rebalance" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.509633 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3603082b-401e-484c-88a3-90c8f67f5824" containerName="swift-ring-rebalance" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.509830 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f67adcdc-2163-48d0-b9d4-013d03f48b5b" containerName="ovn-config" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.509865 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3603082b-401e-484c-88a3-90c8f67f5824" containerName="swift-ring-rebalance" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.510585 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.512540 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5m642" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.518135 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.525410 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g5p9c"] Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.578673 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd49k\" (UniqueName: \"kubernetes.io/projected/abd7c8df-1c06-4098-acc5-8ac65805e879-kube-api-access-nd49k\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.578723 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-db-sync-config-data\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.578823 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-combined-ca-bundle\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.578888 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-config-data\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.579003 4945 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f67adcdc-2163-48d0-b9d4-013d03f48b5b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.579016 4945 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f67adcdc-2163-48d0-b9d4-013d03f48b5b-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.579027 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r69s\" (UniqueName: \"kubernetes.io/projected/f67adcdc-2163-48d0-b9d4-013d03f48b5b-kube-api-access-2r69s\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.636458 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xc2dd-config-mndst" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.636475 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xc2dd-config-mndst" event={"ID":"f67adcdc-2163-48d0-b9d4-013d03f48b5b","Type":"ContainerDied","Data":"ca2fa9000eea4c21e9583b5786e56ca95615fff5b580a46b7bd10170816f387a"} Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.636579 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca2fa9000eea4c21e9583b5786e56ca95615fff5b580a46b7bd10170816f387a" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.638581 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hjh59" event={"ID":"3603082b-401e-484c-88a3-90c8f67f5824","Type":"ContainerDied","Data":"8ed5356b9ec4b2f75d7097d794e6ab0a9d18c0f8ea781e3b6d63644e2953f8cd"} Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.638633 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ed5356b9ec4b2f75d7097d794e6ab0a9d18c0f8ea781e3b6d63644e2953f8cd" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.638748 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hjh59" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.680452 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-config-data\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.680555 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd49k\" (UniqueName: \"kubernetes.io/projected/abd7c8df-1c06-4098-acc5-8ac65805e879-kube-api-access-nd49k\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.680592 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-db-sync-config-data\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.680647 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-combined-ca-bundle\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.684922 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-config-data\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.686036 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-combined-ca-bundle\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.687434 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-db-sync-config-data\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.699298 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd49k\" (UniqueName: \"kubernetes.io/projected/abd7c8df-1c06-4098-acc5-8ac65805e879-kube-api-access-nd49k\") pod \"glance-db-sync-g5p9c\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:36 crc kubenswrapper[4945]: I1206 07:15:36.826984 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g5p9c" Dec 06 07:15:37 crc kubenswrapper[4945]: I1206 07:15:37.427665 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g5p9c"] Dec 06 07:15:37 crc kubenswrapper[4945]: W1206 07:15:37.438047 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabd7c8df_1c06_4098_acc5_8ac65805e879.slice/crio-cf7ba877646907a5fe68c0d03c047254fd8cf838fb9160cdeec1a92294d6d589 WatchSource:0}: Error finding container cf7ba877646907a5fe68c0d03c047254fd8cf838fb9160cdeec1a92294d6d589: Status 404 returned error can't find the container with id cf7ba877646907a5fe68c0d03c047254fd8cf838fb9160cdeec1a92294d6d589 Dec 06 07:15:37 crc kubenswrapper[4945]: I1206 07:15:37.473155 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xc2dd-config-mndst"] Dec 06 07:15:37 crc kubenswrapper[4945]: I1206 07:15:37.480835 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-xc2dd-config-mndst"] Dec 06 07:15:37 crc kubenswrapper[4945]: I1206 07:15:37.647935 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g5p9c" event={"ID":"abd7c8df-1c06-4098-acc5-8ac65805e879","Type":"ContainerStarted","Data":"cf7ba877646907a5fe68c0d03c047254fd8cf838fb9160cdeec1a92294d6d589"} Dec 06 07:15:38 crc kubenswrapper[4945]: I1206 07:15:38.964530 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f67adcdc-2163-48d0-b9d4-013d03f48b5b" path="/var/lib/kubelet/pods/f67adcdc-2163-48d0-b9d4-013d03f48b5b/volumes" Dec 06 07:15:40 crc kubenswrapper[4945]: I1206 07:15:40.978188 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:40 crc kubenswrapper[4945]: I1206 07:15:40.991144 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") pod \"swift-storage-0\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " pod="openstack/swift-storage-0" Dec 06 07:15:41 crc kubenswrapper[4945]: I1206 07:15:41.214706 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 07:15:41 crc kubenswrapper[4945]: I1206 07:15:41.445965 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 07:15:41 crc kubenswrapper[4945]: I1206 07:15:41.753321 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:15:41 crc kubenswrapper[4945]: I1206 07:15:41.991178 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-5d2dk"] Dec 06 07:15:41 crc kubenswrapper[4945]: I1206 07:15:41.992921 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5d2dk" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.000205 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3788-account-create-update-bsdpx"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.001603 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3788-account-create-update-bsdpx" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.006882 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-5d2dk"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.007629 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.014234 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3788-account-create-update-bsdpx"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.120813 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-rbrz6"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.122433 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rbrz6" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.131879 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rbrz6"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.146929 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-4ffa-account-create-update-wqkng"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.148858 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4ffa-account-create-update-wqkng" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.152519 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.161404 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4ffa-account-create-update-wqkng"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.162395 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt25l\" (UniqueName: \"kubernetes.io/projected/8035b30e-21e0-488a-bc9d-78d14f0e11a9-kube-api-access-wt25l\") pod \"cinder-3788-account-create-update-bsdpx\" (UID: \"8035b30e-21e0-488a-bc9d-78d14f0e11a9\") " pod="openstack/cinder-3788-account-create-update-bsdpx" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.162451 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8035b30e-21e0-488a-bc9d-78d14f0e11a9-operator-scripts\") pod \"cinder-3788-account-create-update-bsdpx\" (UID: \"8035b30e-21e0-488a-bc9d-78d14f0e11a9\") " pod="openstack/cinder-3788-account-create-update-bsdpx" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.162483 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6cfx\" (UniqueName: \"kubernetes.io/projected/10ba62ff-07ab-489e-84b6-059d2299e606-kube-api-access-x6cfx\") pod \"barbican-db-create-5d2dk\" (UID: \"10ba62ff-07ab-489e-84b6-059d2299e606\") " pod="openstack/barbican-db-create-5d2dk" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.162766 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10ba62ff-07ab-489e-84b6-059d2299e606-operator-scripts\") pod \"barbican-db-create-5d2dk\" (UID: \"10ba62ff-07ab-489e-84b6-059d2299e606\") " pod="openstack/barbican-db-create-5d2dk" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.264819 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg8xg\" (UniqueName: \"kubernetes.io/projected/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-kube-api-access-lg8xg\") pod \"cinder-db-create-rbrz6\" (UID: \"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729\") " pod="openstack/cinder-db-create-rbrz6" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.264904 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scpz5\" (UniqueName: \"kubernetes.io/projected/dd562125-d6e3-4f47-8bc0-cace5b235a27-kube-api-access-scpz5\") pod \"barbican-4ffa-account-create-update-wqkng\" (UID: \"dd562125-d6e3-4f47-8bc0-cace5b235a27\") " pod="openstack/barbican-4ffa-account-create-update-wqkng" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.264945 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10ba62ff-07ab-489e-84b6-059d2299e606-operator-scripts\") pod \"barbican-db-create-5d2dk\" (UID: \"10ba62ff-07ab-489e-84b6-059d2299e606\") " pod="openstack/barbican-db-create-5d2dk" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.265005 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd562125-d6e3-4f47-8bc0-cace5b235a27-operator-scripts\") pod \"barbican-4ffa-account-create-update-wqkng\" (UID: \"dd562125-d6e3-4f47-8bc0-cace5b235a27\") " pod="openstack/barbican-4ffa-account-create-update-wqkng" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.265359 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt25l\" (UniqueName: \"kubernetes.io/projected/8035b30e-21e0-488a-bc9d-78d14f0e11a9-kube-api-access-wt25l\") pod \"cinder-3788-account-create-update-bsdpx\" (UID: \"8035b30e-21e0-488a-bc9d-78d14f0e11a9\") " pod="openstack/cinder-3788-account-create-update-bsdpx" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.265399 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8035b30e-21e0-488a-bc9d-78d14f0e11a9-operator-scripts\") pod \"cinder-3788-account-create-update-bsdpx\" (UID: \"8035b30e-21e0-488a-bc9d-78d14f0e11a9\") " pod="openstack/cinder-3788-account-create-update-bsdpx" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.265459 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-operator-scripts\") pod \"cinder-db-create-rbrz6\" (UID: \"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729\") " pod="openstack/cinder-db-create-rbrz6" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.265495 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6cfx\" (UniqueName: \"kubernetes.io/projected/10ba62ff-07ab-489e-84b6-059d2299e606-kube-api-access-x6cfx\") pod \"barbican-db-create-5d2dk\" (UID: \"10ba62ff-07ab-489e-84b6-059d2299e606\") " pod="openstack/barbican-db-create-5d2dk" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.266844 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10ba62ff-07ab-489e-84b6-059d2299e606-operator-scripts\") pod \"barbican-db-create-5d2dk\" (UID: \"10ba62ff-07ab-489e-84b6-059d2299e606\") " pod="openstack/barbican-db-create-5d2dk" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.267356 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8035b30e-21e0-488a-bc9d-78d14f0e11a9-operator-scripts\") pod \"cinder-3788-account-create-update-bsdpx\" (UID: \"8035b30e-21e0-488a-bc9d-78d14f0e11a9\") " pod="openstack/cinder-3788-account-create-update-bsdpx" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.291231 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6cfx\" (UniqueName: \"kubernetes.io/projected/10ba62ff-07ab-489e-84b6-059d2299e606-kube-api-access-x6cfx\") pod \"barbican-db-create-5d2dk\" (UID: \"10ba62ff-07ab-489e-84b6-059d2299e606\") " pod="openstack/barbican-db-create-5d2dk" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.294014 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt25l\" (UniqueName: \"kubernetes.io/projected/8035b30e-21e0-488a-bc9d-78d14f0e11a9-kube-api-access-wt25l\") pod \"cinder-3788-account-create-update-bsdpx\" (UID: \"8035b30e-21e0-488a-bc9d-78d14f0e11a9\") " pod="openstack/cinder-3788-account-create-update-bsdpx" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.364428 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5d2dk" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.368951 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg8xg\" (UniqueName: \"kubernetes.io/projected/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-kube-api-access-lg8xg\") pod \"cinder-db-create-rbrz6\" (UID: \"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729\") " pod="openstack/cinder-db-create-rbrz6" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.368999 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scpz5\" (UniqueName: \"kubernetes.io/projected/dd562125-d6e3-4f47-8bc0-cace5b235a27-kube-api-access-scpz5\") pod \"barbican-4ffa-account-create-update-wqkng\" (UID: \"dd562125-d6e3-4f47-8bc0-cace5b235a27\") " pod="openstack/barbican-4ffa-account-create-update-wqkng" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.369049 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd562125-d6e3-4f47-8bc0-cace5b235a27-operator-scripts\") pod \"barbican-4ffa-account-create-update-wqkng\" (UID: \"dd562125-d6e3-4f47-8bc0-cace5b235a27\") " pod="openstack/barbican-4ffa-account-create-update-wqkng" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.369098 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-operator-scripts\") pod \"cinder-db-create-rbrz6\" (UID: \"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729\") " pod="openstack/cinder-db-create-rbrz6" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.369859 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-operator-scripts\") pod \"cinder-db-create-rbrz6\" (UID: \"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729\") " pod="openstack/cinder-db-create-rbrz6" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.370965 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd562125-d6e3-4f47-8bc0-cace5b235a27-operator-scripts\") pod \"barbican-4ffa-account-create-update-wqkng\" (UID: \"dd562125-d6e3-4f47-8bc0-cace5b235a27\") " pod="openstack/barbican-4ffa-account-create-update-wqkng" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.372869 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3788-account-create-update-bsdpx" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.471220 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scpz5\" (UniqueName: \"kubernetes.io/projected/dd562125-d6e3-4f47-8bc0-cace5b235a27-kube-api-access-scpz5\") pod \"barbican-4ffa-account-create-update-wqkng\" (UID: \"dd562125-d6e3-4f47-8bc0-cace5b235a27\") " pod="openstack/barbican-4ffa-account-create-update-wqkng" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.479224 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-975b-account-create-update-rhkgz"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.480334 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-975b-account-create-update-rhkgz" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.481186 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4ffa-account-create-update-wqkng" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.510747 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.512902 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg8xg\" (UniqueName: \"kubernetes.io/projected/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-kube-api-access-lg8xg\") pod \"cinder-db-create-rbrz6\" (UID: \"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729\") " pod="openstack/cinder-db-create-rbrz6" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.525897 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-975b-account-create-update-rhkgz"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.544774 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.559405 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-4vz7h"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.560720 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4vz7h" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.576613 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4vz7h"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.594025 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e61d88c2-c6e2-428e-985d-4d49948f5c0f-operator-scripts\") pod \"neutron-db-create-4vz7h\" (UID: \"e61d88c2-c6e2-428e-985d-4d49948f5c0f\") " pod="openstack/neutron-db-create-4vz7h" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.594125 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/824208f4-ec79-494c-9340-b14d7c7c78b6-operator-scripts\") pod \"neutron-975b-account-create-update-rhkgz\" (UID: \"824208f4-ec79-494c-9340-b14d7c7c78b6\") " pod="openstack/neutron-975b-account-create-update-rhkgz" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.594509 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkfmv\" (UniqueName: \"kubernetes.io/projected/e61d88c2-c6e2-428e-985d-4d49948f5c0f-kube-api-access-kkfmv\") pod \"neutron-db-create-4vz7h\" (UID: \"e61d88c2-c6e2-428e-985d-4d49948f5c0f\") " pod="openstack/neutron-db-create-4vz7h" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.594598 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcdpz\" (UniqueName: \"kubernetes.io/projected/824208f4-ec79-494c-9340-b14d7c7c78b6-kube-api-access-kcdpz\") pod \"neutron-975b-account-create-update-rhkgz\" (UID: \"824208f4-ec79-494c-9340-b14d7c7c78b6\") " pod="openstack/neutron-975b-account-create-update-rhkgz" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.696730 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/824208f4-ec79-494c-9340-b14d7c7c78b6-operator-scripts\") pod \"neutron-975b-account-create-update-rhkgz\" (UID: \"824208f4-ec79-494c-9340-b14d7c7c78b6\") " pod="openstack/neutron-975b-account-create-update-rhkgz" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.698050 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkfmv\" (UniqueName: \"kubernetes.io/projected/e61d88c2-c6e2-428e-985d-4d49948f5c0f-kube-api-access-kkfmv\") pod \"neutron-db-create-4vz7h\" (UID: \"e61d88c2-c6e2-428e-985d-4d49948f5c0f\") " pod="openstack/neutron-db-create-4vz7h" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.698123 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcdpz\" (UniqueName: \"kubernetes.io/projected/824208f4-ec79-494c-9340-b14d7c7c78b6-kube-api-access-kcdpz\") pod \"neutron-975b-account-create-update-rhkgz\" (UID: \"824208f4-ec79-494c-9340-b14d7c7c78b6\") " pod="openstack/neutron-975b-account-create-update-rhkgz" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.698414 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e61d88c2-c6e2-428e-985d-4d49948f5c0f-operator-scripts\") pod \"neutron-db-create-4vz7h\" (UID: \"e61d88c2-c6e2-428e-985d-4d49948f5c0f\") " pod="openstack/neutron-db-create-4vz7h" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.698717 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/824208f4-ec79-494c-9340-b14d7c7c78b6-operator-scripts\") pod \"neutron-975b-account-create-update-rhkgz\" (UID: \"824208f4-ec79-494c-9340-b14d7c7c78b6\") " pod="openstack/neutron-975b-account-create-update-rhkgz" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.699481 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e61d88c2-c6e2-428e-985d-4d49948f5c0f-operator-scripts\") pod \"neutron-db-create-4vz7h\" (UID: \"e61d88c2-c6e2-428e-985d-4d49948f5c0f\") " pod="openstack/neutron-db-create-4vz7h" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.745068 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rbrz6" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.821152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"88e498f32185431bf282d55d7dab63531bc10a134d582784ca5237d4afa2b48f"} Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.826499 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkfmv\" (UniqueName: \"kubernetes.io/projected/e61d88c2-c6e2-428e-985d-4d49948f5c0f-kube-api-access-kkfmv\") pod \"neutron-db-create-4vz7h\" (UID: \"e61d88c2-c6e2-428e-985d-4d49948f5c0f\") " pod="openstack/neutron-db-create-4vz7h" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.832327 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcdpz\" (UniqueName: \"kubernetes.io/projected/824208f4-ec79-494c-9340-b14d7c7c78b6-kube-api-access-kcdpz\") pod \"neutron-975b-account-create-update-rhkgz\" (UID: \"824208f4-ec79-494c-9340-b14d7c7c78b6\") " pod="openstack/neutron-975b-account-create-update-rhkgz" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.837259 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-975b-account-create-update-rhkgz" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.861482 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-r6cs4"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.869125 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-r6cs4"] Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.869294 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.875430 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.875690 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.875765 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wf8d7" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.875843 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 07:15:42 crc kubenswrapper[4945]: I1206 07:15:42.972379 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4vz7h" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.030031 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-config-data\") pod \"keystone-db-sync-r6cs4\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.030112 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-combined-ca-bundle\") pod \"keystone-db-sync-r6cs4\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.032358 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khbln\" (UniqueName: \"kubernetes.io/projected/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-kube-api-access-khbln\") pod \"keystone-db-sync-r6cs4\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.134837 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-config-data\") pod \"keystone-db-sync-r6cs4\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.134938 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-combined-ca-bundle\") pod \"keystone-db-sync-r6cs4\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.135020 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khbln\" (UniqueName: \"kubernetes.io/projected/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-kube-api-access-khbln\") pod \"keystone-db-sync-r6cs4\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.141960 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-config-data\") pod \"keystone-db-sync-r6cs4\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.155041 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-combined-ca-bundle\") pod \"keystone-db-sync-r6cs4\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.168039 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-5d2dk"] Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.168955 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khbln\" (UniqueName: \"kubernetes.io/projected/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-kube-api-access-khbln\") pod \"keystone-db-sync-r6cs4\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.237689 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.261415 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3788-account-create-update-bsdpx"] Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.578828 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4ffa-account-create-update-wqkng"] Dec 06 07:15:43 crc kubenswrapper[4945]: W1206 07:15:43.591136 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd562125_d6e3_4f47_8bc0_cace5b235a27.slice/crio-438a7cf4f19dba0f3f7e9233f45d6dcb17ece6cfd59001dbdebcd1166dba5b1f WatchSource:0}: Error finding container 438a7cf4f19dba0f3f7e9233f45d6dcb17ece6cfd59001dbdebcd1166dba5b1f: Status 404 returned error can't find the container with id 438a7cf4f19dba0f3f7e9233f45d6dcb17ece6cfd59001dbdebcd1166dba5b1f Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.673306 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rbrz6"] Dec 06 07:15:43 crc kubenswrapper[4945]: W1206 07:15:43.729160 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda35c1b0f_f7ed_47ea_b3b8_dc855c43a729.slice/crio-d298a1acfcef328560cea6bfd72d79571c9922931678f93feac8a79e04bbba8c WatchSource:0}: Error finding container d298a1acfcef328560cea6bfd72d79571c9922931678f93feac8a79e04bbba8c: Status 404 returned error can't find the container with id d298a1acfcef328560cea6bfd72d79571c9922931678f93feac8a79e04bbba8c Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.751624 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-975b-account-create-update-rhkgz"] Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.767737 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4vz7h"] Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.838749 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5d2dk" event={"ID":"10ba62ff-07ab-489e-84b6-059d2299e606","Type":"ContainerStarted","Data":"64091da47e5612c87b8350905f3c5270e29715775372b5e9af40732fcacee0da"} Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.838812 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5d2dk" event={"ID":"10ba62ff-07ab-489e-84b6-059d2299e606","Type":"ContainerStarted","Data":"6e42cf004625cfbe944c61cfd00ab892ae5cafe32d1e269f02383e339a8e365f"} Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.841764 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4vz7h" event={"ID":"e61d88c2-c6e2-428e-985d-4d49948f5c0f","Type":"ContainerStarted","Data":"4207778b5ea1450614e915df8dfa221e856151170b6980814407c3efcd6e13be"} Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.846705 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-975b-account-create-update-rhkgz" event={"ID":"824208f4-ec79-494c-9340-b14d7c7c78b6","Type":"ContainerStarted","Data":"fac47ea0dabb0f74b3b39c8383d4df22d417b6ad2b9c9277375ba8f411bb76fd"} Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.862561 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-5d2dk" podStartSLOduration=2.862539382 podStartE2EDuration="2.862539382s" podCreationTimestamp="2025-12-06 07:15:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:43.854893145 +0000 UTC m=+1357.309754179" watchObservedRunningTime="2025-12-06 07:15:43.862539382 +0000 UTC m=+1357.317400426" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.864107 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3788-account-create-update-bsdpx" event={"ID":"8035b30e-21e0-488a-bc9d-78d14f0e11a9","Type":"ContainerStarted","Data":"e01ace9d4e5e9d743ae9af7802df7c36ec09040535d59f26854b256f8a64cc9b"} Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.864156 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3788-account-create-update-bsdpx" event={"ID":"8035b30e-21e0-488a-bc9d-78d14f0e11a9","Type":"ContainerStarted","Data":"e99e3454be4c2b57c3d631800813da69894b8d498ab8f74de220ff635b28253d"} Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.887197 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-3788-account-create-update-bsdpx" podStartSLOduration=2.887172688 podStartE2EDuration="2.887172688s" podCreationTimestamp="2025-12-06 07:15:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:15:43.883478558 +0000 UTC m=+1357.338339602" watchObservedRunningTime="2025-12-06 07:15:43.887172688 +0000 UTC m=+1357.342033732" Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.893560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4ffa-account-create-update-wqkng" event={"ID":"dd562125-d6e3-4f47-8bc0-cace5b235a27","Type":"ContainerStarted","Data":"438a7cf4f19dba0f3f7e9233f45d6dcb17ece6cfd59001dbdebcd1166dba5b1f"} Dec 06 07:15:43 crc kubenswrapper[4945]: I1206 07:15:43.945569 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rbrz6" event={"ID":"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729","Type":"ContainerStarted","Data":"d298a1acfcef328560cea6bfd72d79571c9922931678f93feac8a79e04bbba8c"} Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.004631 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-r6cs4"] Dec 06 07:15:44 crc kubenswrapper[4945]: W1206 07:15:44.013128 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a5cd6b1_5645_4e0a_be9d_aed6f7930c53.slice/crio-03710edf61147df9e1865c0745ca53b0f49a15fb3e772665fa01509a4ee525c0 WatchSource:0}: Error finding container 03710edf61147df9e1865c0745ca53b0f49a15fb3e772665fa01509a4ee525c0: Status 404 returned error can't find the container with id 03710edf61147df9e1865c0745ca53b0f49a15fb3e772665fa01509a4ee525c0 Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.964799 4945 generic.go:334] "Generic (PLEG): container finished" podID="e61d88c2-c6e2-428e-985d-4d49948f5c0f" containerID="b300df8609ad712283eee0541bfa67913ab230d553149ead1b538a7e5cf6a1d0" exitCode=0 Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.970801 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4vz7h" event={"ID":"e61d88c2-c6e2-428e-985d-4d49948f5c0f","Type":"ContainerDied","Data":"b300df8609ad712283eee0541bfa67913ab230d553149ead1b538a7e5cf6a1d0"} Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.974178 4945 generic.go:334] "Generic (PLEG): container finished" podID="824208f4-ec79-494c-9340-b14d7c7c78b6" containerID="1bfccd71aa9901558b2bdd7c5eb7435e28b26696e7097a7b152a71a444c574b1" exitCode=0 Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.974236 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-975b-account-create-update-rhkgz" event={"ID":"824208f4-ec79-494c-9340-b14d7c7c78b6","Type":"ContainerDied","Data":"1bfccd71aa9901558b2bdd7c5eb7435e28b26696e7097a7b152a71a444c574b1"} Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.976434 4945 generic.go:334] "Generic (PLEG): container finished" podID="8035b30e-21e0-488a-bc9d-78d14f0e11a9" containerID="e01ace9d4e5e9d743ae9af7802df7c36ec09040535d59f26854b256f8a64cc9b" exitCode=0 Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.976475 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3788-account-create-update-bsdpx" event={"ID":"8035b30e-21e0-488a-bc9d-78d14f0e11a9","Type":"ContainerDied","Data":"e01ace9d4e5e9d743ae9af7802df7c36ec09040535d59f26854b256f8a64cc9b"} Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.978559 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-r6cs4" event={"ID":"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53","Type":"ContainerStarted","Data":"03710edf61147df9e1865c0745ca53b0f49a15fb3e772665fa01509a4ee525c0"} Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.992853 4945 generic.go:334] "Generic (PLEG): container finished" podID="dd562125-d6e3-4f47-8bc0-cace5b235a27" containerID="0503309da1afde1f127ec513bd61decd36e94ee9c302693c769b55d708825aeb" exitCode=0 Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.993013 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4ffa-account-create-update-wqkng" event={"ID":"dd562125-d6e3-4f47-8bc0-cace5b235a27","Type":"ContainerDied","Data":"0503309da1afde1f127ec513bd61decd36e94ee9c302693c769b55d708825aeb"} Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.995743 4945 generic.go:334] "Generic (PLEG): container finished" podID="a35c1b0f-f7ed-47ea-b3b8-dc855c43a729" containerID="29cdfb50edd3a43da06ed0091aa1d96671bf431d904eb9c175dff12ba56e0477" exitCode=0 Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.995805 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rbrz6" event={"ID":"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729","Type":"ContainerDied","Data":"29cdfb50edd3a43da06ed0091aa1d96671bf431d904eb9c175dff12ba56e0477"} Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.997088 4945 generic.go:334] "Generic (PLEG): container finished" podID="10ba62ff-07ab-489e-84b6-059d2299e606" containerID="64091da47e5612c87b8350905f3c5270e29715775372b5e9af40732fcacee0da" exitCode=0 Dec 06 07:15:44 crc kubenswrapper[4945]: I1206 07:15:44.997113 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5d2dk" event={"ID":"10ba62ff-07ab-489e-84b6-059d2299e606","Type":"ContainerDied","Data":"64091da47e5612c87b8350905f3c5270e29715775372b5e9af40732fcacee0da"} Dec 06 07:15:46 crc kubenswrapper[4945]: I1206 07:15:46.286083 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"57a847dca15d392b38d22b41bbbc40b779272394f028698e226e315d75751d47"} Dec 06 07:15:46 crc kubenswrapper[4945]: I1206 07:15:46.286434 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"8f302203e6126bf00e33188ea6dd1f7e034a8cbfe7c5182693c187c7f24b819a"} Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.512867 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-975b-account-create-update-rhkgz" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.601553 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcdpz\" (UniqueName: \"kubernetes.io/projected/824208f4-ec79-494c-9340-b14d7c7c78b6-kube-api-access-kcdpz\") pod \"824208f4-ec79-494c-9340-b14d7c7c78b6\" (UID: \"824208f4-ec79-494c-9340-b14d7c7c78b6\") " Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.601802 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/824208f4-ec79-494c-9340-b14d7c7c78b6-operator-scripts\") pod \"824208f4-ec79-494c-9340-b14d7c7c78b6\" (UID: \"824208f4-ec79-494c-9340-b14d7c7c78b6\") " Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.603242 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/824208f4-ec79-494c-9340-b14d7c7c78b6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "824208f4-ec79-494c-9340-b14d7c7c78b6" (UID: "824208f4-ec79-494c-9340-b14d7c7c78b6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.627674 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/824208f4-ec79-494c-9340-b14d7c7c78b6-kube-api-access-kcdpz" (OuterVolumeSpecName: "kube-api-access-kcdpz") pod "824208f4-ec79-494c-9340-b14d7c7c78b6" (UID: "824208f4-ec79-494c-9340-b14d7c7c78b6"). InnerVolumeSpecName "kube-api-access-kcdpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.703780 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/824208f4-ec79-494c-9340-b14d7c7c78b6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.703819 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcdpz\" (UniqueName: \"kubernetes.io/projected/824208f4-ec79-494c-9340-b14d7c7c78b6-kube-api-access-kcdpz\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.779626 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4vz7h" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.787999 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5d2dk" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.807811 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4ffa-account-create-update-wqkng" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.810141 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6cfx\" (UniqueName: \"kubernetes.io/projected/10ba62ff-07ab-489e-84b6-059d2299e606-kube-api-access-x6cfx\") pod \"10ba62ff-07ab-489e-84b6-059d2299e606\" (UID: \"10ba62ff-07ab-489e-84b6-059d2299e606\") " Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.810237 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10ba62ff-07ab-489e-84b6-059d2299e606-operator-scripts\") pod \"10ba62ff-07ab-489e-84b6-059d2299e606\" (UID: \"10ba62ff-07ab-489e-84b6-059d2299e606\") " Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.810357 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkfmv\" (UniqueName: \"kubernetes.io/projected/e61d88c2-c6e2-428e-985d-4d49948f5c0f-kube-api-access-kkfmv\") pod \"e61d88c2-c6e2-428e-985d-4d49948f5c0f\" (UID: \"e61d88c2-c6e2-428e-985d-4d49948f5c0f\") " Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.810432 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e61d88c2-c6e2-428e-985d-4d49948f5c0f-operator-scripts\") pod \"e61d88c2-c6e2-428e-985d-4d49948f5c0f\" (UID: \"e61d88c2-c6e2-428e-985d-4d49948f5c0f\") " Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.813228 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e61d88c2-c6e2-428e-985d-4d49948f5c0f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e61d88c2-c6e2-428e-985d-4d49948f5c0f" (UID: "e61d88c2-c6e2-428e-985d-4d49948f5c0f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.816966 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3788-account-create-update-bsdpx" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.818203 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e61d88c2-c6e2-428e-985d-4d49948f5c0f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.824788 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10ba62ff-07ab-489e-84b6-059d2299e606-kube-api-access-x6cfx" (OuterVolumeSpecName: "kube-api-access-x6cfx") pod "10ba62ff-07ab-489e-84b6-059d2299e606" (UID: "10ba62ff-07ab-489e-84b6-059d2299e606"). InnerVolumeSpecName "kube-api-access-x6cfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.834098 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e61d88c2-c6e2-428e-985d-4d49948f5c0f-kube-api-access-kkfmv" (OuterVolumeSpecName: "kube-api-access-kkfmv") pod "e61d88c2-c6e2-428e-985d-4d49948f5c0f" (UID: "e61d88c2-c6e2-428e-985d-4d49948f5c0f"). InnerVolumeSpecName "kube-api-access-kkfmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.919646 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scpz5\" (UniqueName: \"kubernetes.io/projected/dd562125-d6e3-4f47-8bc0-cace5b235a27-kube-api-access-scpz5\") pod \"dd562125-d6e3-4f47-8bc0-cace5b235a27\" (UID: \"dd562125-d6e3-4f47-8bc0-cace5b235a27\") " Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.919734 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt25l\" (UniqueName: \"kubernetes.io/projected/8035b30e-21e0-488a-bc9d-78d14f0e11a9-kube-api-access-wt25l\") pod \"8035b30e-21e0-488a-bc9d-78d14f0e11a9\" (UID: \"8035b30e-21e0-488a-bc9d-78d14f0e11a9\") " Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.919969 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8035b30e-21e0-488a-bc9d-78d14f0e11a9-operator-scripts\") pod \"8035b30e-21e0-488a-bc9d-78d14f0e11a9\" (UID: \"8035b30e-21e0-488a-bc9d-78d14f0e11a9\") " Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.920014 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd562125-d6e3-4f47-8bc0-cace5b235a27-operator-scripts\") pod \"dd562125-d6e3-4f47-8bc0-cace5b235a27\" (UID: \"dd562125-d6e3-4f47-8bc0-cace5b235a27\") " Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.920382 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6cfx\" (UniqueName: \"kubernetes.io/projected/10ba62ff-07ab-489e-84b6-059d2299e606-kube-api-access-x6cfx\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.920393 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkfmv\" (UniqueName: \"kubernetes.io/projected/e61d88c2-c6e2-428e-985d-4d49948f5c0f-kube-api-access-kkfmv\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.921425 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd562125-d6e3-4f47-8bc0-cace5b235a27-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd562125-d6e3-4f47-8bc0-cace5b235a27" (UID: "dd562125-d6e3-4f47-8bc0-cace5b235a27"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.921530 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8035b30e-21e0-488a-bc9d-78d14f0e11a9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8035b30e-21e0-488a-bc9d-78d14f0e11a9" (UID: "8035b30e-21e0-488a-bc9d-78d14f0e11a9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.924413 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8035b30e-21e0-488a-bc9d-78d14f0e11a9-kube-api-access-wt25l" (OuterVolumeSpecName: "kube-api-access-wt25l") pod "8035b30e-21e0-488a-bc9d-78d14f0e11a9" (UID: "8035b30e-21e0-488a-bc9d-78d14f0e11a9"). InnerVolumeSpecName "kube-api-access-wt25l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.925199 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd562125-d6e3-4f47-8bc0-cace5b235a27-kube-api-access-scpz5" (OuterVolumeSpecName: "kube-api-access-scpz5") pod "dd562125-d6e3-4f47-8bc0-cace5b235a27" (UID: "dd562125-d6e3-4f47-8bc0-cace5b235a27"). InnerVolumeSpecName "kube-api-access-scpz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:47 crc kubenswrapper[4945]: I1206 07:15:47.949474 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rbrz6" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.021686 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lg8xg\" (UniqueName: \"kubernetes.io/projected/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-kube-api-access-lg8xg\") pod \"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729\" (UID: \"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729\") " Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.021749 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-operator-scripts\") pod \"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729\" (UID: \"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729\") " Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.022313 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a35c1b0f-f7ed-47ea-b3b8-dc855c43a729" (UID: "a35c1b0f-f7ed-47ea-b3b8-dc855c43a729"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.023163 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8035b30e-21e0-488a-bc9d-78d14f0e11a9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.023186 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd562125-d6e3-4f47-8bc0-cace5b235a27-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.023198 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scpz5\" (UniqueName: \"kubernetes.io/projected/dd562125-d6e3-4f47-8bc0-cace5b235a27-kube-api-access-scpz5\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.023214 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt25l\" (UniqueName: \"kubernetes.io/projected/8035b30e-21e0-488a-bc9d-78d14f0e11a9-kube-api-access-wt25l\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.023228 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.035569 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-kube-api-access-lg8xg" (OuterVolumeSpecName: "kube-api-access-lg8xg") pod "a35c1b0f-f7ed-47ea-b3b8-dc855c43a729" (UID: "a35c1b0f-f7ed-47ea-b3b8-dc855c43a729"). InnerVolumeSpecName "kube-api-access-lg8xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.124494 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lg8xg\" (UniqueName: \"kubernetes.io/projected/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729-kube-api-access-lg8xg\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.310864 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3788-account-create-update-bsdpx" event={"ID":"8035b30e-21e0-488a-bc9d-78d14f0e11a9","Type":"ContainerDied","Data":"e99e3454be4c2b57c3d631800813da69894b8d498ab8f74de220ff635b28253d"} Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.310930 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e99e3454be4c2b57c3d631800813da69894b8d498ab8f74de220ff635b28253d" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.310893 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3788-account-create-update-bsdpx" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.312812 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4ffa-account-create-update-wqkng" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.313030 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4ffa-account-create-update-wqkng" event={"ID":"dd562125-d6e3-4f47-8bc0-cace5b235a27","Type":"ContainerDied","Data":"438a7cf4f19dba0f3f7e9233f45d6dcb17ece6cfd59001dbdebcd1166dba5b1f"} Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.313097 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="438a7cf4f19dba0f3f7e9233f45d6dcb17ece6cfd59001dbdebcd1166dba5b1f" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.315178 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rbrz6" event={"ID":"a35c1b0f-f7ed-47ea-b3b8-dc855c43a729","Type":"ContainerDied","Data":"d298a1acfcef328560cea6bfd72d79571c9922931678f93feac8a79e04bbba8c"} Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.315196 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rbrz6" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.315207 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d298a1acfcef328560cea6bfd72d79571c9922931678f93feac8a79e04bbba8c" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.320189 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5d2dk" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.320229 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5d2dk" event={"ID":"10ba62ff-07ab-489e-84b6-059d2299e606","Type":"ContainerDied","Data":"6e42cf004625cfbe944c61cfd00ab892ae5cafe32d1e269f02383e339a8e365f"} Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.320309 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e42cf004625cfbe944c61cfd00ab892ae5cafe32d1e269f02383e339a8e365f" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.322375 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4vz7h" event={"ID":"e61d88c2-c6e2-428e-985d-4d49948f5c0f","Type":"ContainerDied","Data":"4207778b5ea1450614e915df8dfa221e856151170b6980814407c3efcd6e13be"} Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.322417 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4207778b5ea1450614e915df8dfa221e856151170b6980814407c3efcd6e13be" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.322483 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4vz7h" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.325832 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-975b-account-create-update-rhkgz" event={"ID":"824208f4-ec79-494c-9340-b14d7c7c78b6","Type":"ContainerDied","Data":"fac47ea0dabb0f74b3b39c8383d4df22d417b6ad2b9c9277375ba8f411bb76fd"} Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.325868 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fac47ea0dabb0f74b3b39c8383d4df22d417b6ad2b9c9277375ba8f411bb76fd" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.325973 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-975b-account-create-update-rhkgz" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.352110 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10ba62ff-07ab-489e-84b6-059d2299e606-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "10ba62ff-07ab-489e-84b6-059d2299e606" (UID: "10ba62ff-07ab-489e-84b6-059d2299e606"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:48 crc kubenswrapper[4945]: I1206 07:15:48.430919 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10ba62ff-07ab-489e-84b6-059d2299e606-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:49 crc kubenswrapper[4945]: I1206 07:15:49.335657 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"ff9b25f9862ebd3ee6236d1ae7fa95a0a44adbb1a4d9be0d4f84005b7298f8fd"} Dec 06 07:15:50 crc kubenswrapper[4945]: I1206 07:15:50.353428 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"a0fcd13fd05e6559ba8dfeed754e555d93dc6f54fcefdb814974d89ffc3bac43"} Dec 06 07:16:00 crc kubenswrapper[4945]: E1206 07:16:00.511095 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63" Dec 06 07:16:00 crc kubenswrapper[4945]: E1206 07:16:00.511893 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nd49k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-g5p9c_openstack(abd7c8df-1c06-4098-acc5-8ac65805e879): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:16:00 crc kubenswrapper[4945]: E1206 07:16:00.513433 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-g5p9c" podUID="abd7c8df-1c06-4098-acc5-8ac65805e879" Dec 06 07:16:01 crc kubenswrapper[4945]: E1206 07:16:01.459099 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63\\\"\"" pod="openstack/glance-db-sync-g5p9c" podUID="abd7c8df-1c06-4098-acc5-8ac65805e879" Dec 06 07:16:03 crc kubenswrapper[4945]: E1206 07:16:03.049437 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-keystone@sha256:a7b6fa2f16a882674624b48939737e2bd95da7bef60db593a8e6e4d397fa516c" Dec 06 07:16:03 crc kubenswrapper[4945]: E1206 07:16:03.049915 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.io/podified-antelope-centos9/openstack-keystone@sha256:a7b6fa2f16a882674624b48939737e2bd95da7bef60db593a8e6e4d397fa516c,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-khbln,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-r6cs4_openstack(9a5cd6b1-5645-4e0a-be9d-aed6f7930c53): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:16:03 crc kubenswrapper[4945]: E1206 07:16:03.051132 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/keystone-db-sync-r6cs4" podUID="9a5cd6b1-5645-4e0a-be9d-aed6f7930c53" Dec 06 07:16:03 crc kubenswrapper[4945]: E1206 07:16:03.475720 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-keystone@sha256:a7b6fa2f16a882674624b48939737e2bd95da7bef60db593a8e6e4d397fa516c\\\"\"" pod="openstack/keystone-db-sync-r6cs4" podUID="9a5cd6b1-5645-4e0a-be9d-aed6f7930c53" Dec 06 07:16:06 crc kubenswrapper[4945]: I1206 07:16:06.503308 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"1991abb13d702f2e5d57933ab6d10cf0d4c07b1134ee92235420a55b874e1f9e"} Dec 06 07:16:06 crc kubenswrapper[4945]: I1206 07:16:06.503907 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"cbfe05d94d7270d286fed8b4840eccd1701ac59c60fb2b3214890ceec204a5cd"} Dec 06 07:16:06 crc kubenswrapper[4945]: I1206 07:16:06.503926 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"195d75cef2ef89afdb1df93e86f8e6e38bae58c3938182ede8e3b5daa3fde195"} Dec 06 07:16:07 crc kubenswrapper[4945]: I1206 07:16:07.521145 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"de0927bc9e6dd18e1a79f623e51e8901bc80dec257144ca686e887a312f70a86"} Dec 06 07:16:08 crc kubenswrapper[4945]: I1206 07:16:08.535086 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"04e71c5a4df960ad72d6476b9f52dc9c05815786d261a2816740f83faab2d771"} Dec 06 07:16:09 crc kubenswrapper[4945]: I1206 07:16:09.550213 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"e0fbc46abb12d20db2057256ae54a1bd2c5b3df4bd87a0d9a1700ac12c2ce2c1"} Dec 06 07:16:09 crc kubenswrapper[4945]: I1206 07:16:09.550637 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"0c80efac78458edb45ee60a60f64b441b635ee4a49351c959c569def27d0e665"} Dec 06 07:16:09 crc kubenswrapper[4945]: I1206 07:16:09.550680 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"a8901783d5561a470b55820422085e5ac6d8ccd2649a4eec8701a774fe804faf"} Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.567409 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"789049d9004e9570b8c0301ec30ded93adc7a99fef755791d3e1828734190e19"} Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.567911 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"ff4e063c23aaff06ab010a43b39361561a092ed784e7fd983facb5b882fd3225"} Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.567937 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerStarted","Data":"6088cfa6a692765f885d01e25babb5993fb4320f6715c412196029c872327c41"} Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.613809 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.925390933 podStartE2EDuration="1m3.613783817s" podCreationTimestamp="2025-12-06 07:15:07 +0000 UTC" firstStartedPulling="2025-12-06 07:15:42.561470973 +0000 UTC m=+1356.016332017" lastFinishedPulling="2025-12-06 07:16:08.249863847 +0000 UTC m=+1381.704724901" observedRunningTime="2025-12-06 07:16:10.606839388 +0000 UTC m=+1384.061700462" watchObservedRunningTime="2025-12-06 07:16:10.613783817 +0000 UTC m=+1384.068644861" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.946615 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-nhnm5"] Dec 06 07:16:10 crc kubenswrapper[4945]: E1206 07:16:10.947019 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8035b30e-21e0-488a-bc9d-78d14f0e11a9" containerName="mariadb-account-create-update" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947036 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8035b30e-21e0-488a-bc9d-78d14f0e11a9" containerName="mariadb-account-create-update" Dec 06 07:16:10 crc kubenswrapper[4945]: E1206 07:16:10.947051 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824208f4-ec79-494c-9340-b14d7c7c78b6" containerName="mariadb-account-create-update" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947057 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="824208f4-ec79-494c-9340-b14d7c7c78b6" containerName="mariadb-account-create-update" Dec 06 07:16:10 crc kubenswrapper[4945]: E1206 07:16:10.947073 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e61d88c2-c6e2-428e-985d-4d49948f5c0f" containerName="mariadb-database-create" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947083 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e61d88c2-c6e2-428e-985d-4d49948f5c0f" containerName="mariadb-database-create" Dec 06 07:16:10 crc kubenswrapper[4945]: E1206 07:16:10.947092 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35c1b0f-f7ed-47ea-b3b8-dc855c43a729" containerName="mariadb-database-create" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947097 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35c1b0f-f7ed-47ea-b3b8-dc855c43a729" containerName="mariadb-database-create" Dec 06 07:16:10 crc kubenswrapper[4945]: E1206 07:16:10.947114 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ba62ff-07ab-489e-84b6-059d2299e606" containerName="mariadb-database-create" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947123 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ba62ff-07ab-489e-84b6-059d2299e606" containerName="mariadb-database-create" Dec 06 07:16:10 crc kubenswrapper[4945]: E1206 07:16:10.947140 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd562125-d6e3-4f47-8bc0-cace5b235a27" containerName="mariadb-account-create-update" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947151 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd562125-d6e3-4f47-8bc0-cace5b235a27" containerName="mariadb-account-create-update" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947403 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8035b30e-21e0-488a-bc9d-78d14f0e11a9" containerName="mariadb-account-create-update" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947440 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="824208f4-ec79-494c-9340-b14d7c7c78b6" containerName="mariadb-account-create-update" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947485 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a35c1b0f-f7ed-47ea-b3b8-dc855c43a729" containerName="mariadb-database-create" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947507 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="10ba62ff-07ab-489e-84b6-059d2299e606" containerName="mariadb-database-create" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947533 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e61d88c2-c6e2-428e-985d-4d49948f5c0f" containerName="mariadb-database-create" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.947556 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd562125-d6e3-4f47-8bc0-cace5b235a27" containerName="mariadb-account-create-update" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.948611 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.953904 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.960610 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-config\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.960824 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-swift-storage-0\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.960890 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-sb\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.960921 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc8rn\" (UniqueName: \"kubernetes.io/projected/a68a3b24-6715-478e-9b85-edcc08b905dd-kube-api-access-dc8rn\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.961062 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-nb\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.961196 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-svc\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:10 crc kubenswrapper[4945]: I1206 07:16:10.971459 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-nhnm5"] Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.062791 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-swift-storage-0\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.062860 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-sb\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.062878 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc8rn\" (UniqueName: \"kubernetes.io/projected/a68a3b24-6715-478e-9b85-edcc08b905dd-kube-api-access-dc8rn\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.062929 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-nb\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.062962 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-svc\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.063016 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-config\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.064132 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-svc\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.064189 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-swift-storage-0\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.064339 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-sb\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.064584 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-nb\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.064650 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-config\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.093604 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc8rn\" (UniqueName: \"kubernetes.io/projected/a68a3b24-6715-478e-9b85-edcc08b905dd-kube-api-access-dc8rn\") pod \"dnsmasq-dns-779c5847bc-nhnm5\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.268681 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:11 crc kubenswrapper[4945]: I1206 07:16:11.638011 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-nhnm5"] Dec 06 07:16:12 crc kubenswrapper[4945]: I1206 07:16:12.624198 4945 generic.go:334] "Generic (PLEG): container finished" podID="a68a3b24-6715-478e-9b85-edcc08b905dd" containerID="1edbf1ee4c10c69b75fab6dbda0dd63e28ec920b8c203e295ec3bd09380500a5" exitCode=0 Dec 06 07:16:12 crc kubenswrapper[4945]: I1206 07:16:12.624347 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" event={"ID":"a68a3b24-6715-478e-9b85-edcc08b905dd","Type":"ContainerDied","Data":"1edbf1ee4c10c69b75fab6dbda0dd63e28ec920b8c203e295ec3bd09380500a5"} Dec 06 07:16:12 crc kubenswrapper[4945]: I1206 07:16:12.624546 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" event={"ID":"a68a3b24-6715-478e-9b85-edcc08b905dd","Type":"ContainerStarted","Data":"789bcedca791e94cd1bba82384dcbcaacdf868936c9a7c5cb244e268104c5557"} Dec 06 07:16:13 crc kubenswrapper[4945]: I1206 07:16:13.638453 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" event={"ID":"a68a3b24-6715-478e-9b85-edcc08b905dd","Type":"ContainerStarted","Data":"13de05e3688414d4b0922e8767a080a6d6be044b6a8e10519316924b09fc2d99"} Dec 06 07:16:13 crc kubenswrapper[4945]: I1206 07:16:13.640145 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:13 crc kubenswrapper[4945]: I1206 07:16:13.673782 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" podStartSLOduration=3.673759956 podStartE2EDuration="3.673759956s" podCreationTimestamp="2025-12-06 07:16:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:16:13.665479331 +0000 UTC m=+1387.120340365" watchObservedRunningTime="2025-12-06 07:16:13.673759956 +0000 UTC m=+1387.128621000" Dec 06 07:16:15 crc kubenswrapper[4945]: I1206 07:16:15.668767 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-r6cs4" event={"ID":"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53","Type":"ContainerStarted","Data":"fa3756ffcf3445aad0a9657ca299b426c918cc8325f6302eb842204e4e1fb37a"} Dec 06 07:16:15 crc kubenswrapper[4945]: I1206 07:16:15.670994 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g5p9c" event={"ID":"abd7c8df-1c06-4098-acc5-8ac65805e879","Type":"ContainerStarted","Data":"948373ff86352f7383fc745821f8ce1cfcddc2703574fde9f0f77d1c9931317f"} Dec 06 07:16:15 crc kubenswrapper[4945]: I1206 07:16:15.690420 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-r6cs4" podStartSLOduration=2.315494181 podStartE2EDuration="33.690401398s" podCreationTimestamp="2025-12-06 07:15:42 +0000 UTC" firstStartedPulling="2025-12-06 07:15:44.021154192 +0000 UTC m=+1357.476015236" lastFinishedPulling="2025-12-06 07:16:15.396061419 +0000 UTC m=+1388.850922453" observedRunningTime="2025-12-06 07:16:15.686038819 +0000 UTC m=+1389.140899863" watchObservedRunningTime="2025-12-06 07:16:15.690401398 +0000 UTC m=+1389.145262442" Dec 06 07:16:15 crc kubenswrapper[4945]: I1206 07:16:15.711486 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-g5p9c" podStartSLOduration=2.787051368 podStartE2EDuration="39.711469821s" podCreationTimestamp="2025-12-06 07:15:36 +0000 UTC" firstStartedPulling="2025-12-06 07:15:37.444772415 +0000 UTC m=+1350.899633459" lastFinishedPulling="2025-12-06 07:16:14.369190868 +0000 UTC m=+1387.824051912" observedRunningTime="2025-12-06 07:16:15.70406745 +0000 UTC m=+1389.158928504" watchObservedRunningTime="2025-12-06 07:16:15.711469821 +0000 UTC m=+1389.166330865" Dec 06 07:16:19 crc kubenswrapper[4945]: I1206 07:16:19.705972 4945 generic.go:334] "Generic (PLEG): container finished" podID="9a5cd6b1-5645-4e0a-be9d-aed6f7930c53" containerID="fa3756ffcf3445aad0a9657ca299b426c918cc8325f6302eb842204e4e1fb37a" exitCode=0 Dec 06 07:16:19 crc kubenswrapper[4945]: I1206 07:16:19.706059 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-r6cs4" event={"ID":"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53","Type":"ContainerDied","Data":"fa3756ffcf3445aad0a9657ca299b426c918cc8325f6302eb842204e4e1fb37a"} Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.013788 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.153433 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-config-data\") pod \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.153879 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khbln\" (UniqueName: \"kubernetes.io/projected/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-kube-api-access-khbln\") pod \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.153993 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-combined-ca-bundle\") pod \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\" (UID: \"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53\") " Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.161656 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-kube-api-access-khbln" (OuterVolumeSpecName: "kube-api-access-khbln") pod "9a5cd6b1-5645-4e0a-be9d-aed6f7930c53" (UID: "9a5cd6b1-5645-4e0a-be9d-aed6f7930c53"). InnerVolumeSpecName "kube-api-access-khbln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.186503 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a5cd6b1-5645-4e0a-be9d-aed6f7930c53" (UID: "9a5cd6b1-5645-4e0a-be9d-aed6f7930c53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.230610 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-config-data" (OuterVolumeSpecName: "config-data") pod "9a5cd6b1-5645-4e0a-be9d-aed6f7930c53" (UID: "9a5cd6b1-5645-4e0a-be9d-aed6f7930c53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.256024 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.256058 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khbln\" (UniqueName: \"kubernetes.io/projected/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-kube-api-access-khbln\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.256070 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.270577 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.330573 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-mlclv"] Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.330821 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" podUID="0f520e1c-bdfd-4fc9-9608-237d44a73622" containerName="dnsmasq-dns" containerID="cri-o://877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937" gracePeriod=10 Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.727569 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.728185 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-r6cs4" event={"ID":"9a5cd6b1-5645-4e0a-be9d-aed6f7930c53","Type":"ContainerDied","Data":"03710edf61147df9e1865c0745ca53b0f49a15fb3e772665fa01509a4ee525c0"} Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.728227 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03710edf61147df9e1865c0745ca53b0f49a15fb3e772665fa01509a4ee525c0" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.728299 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-r6cs4" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.731006 4945 generic.go:334] "Generic (PLEG): container finished" podID="0f520e1c-bdfd-4fc9-9608-237d44a73622" containerID="877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937" exitCode=0 Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.731054 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" event={"ID":"0f520e1c-bdfd-4fc9-9608-237d44a73622","Type":"ContainerDied","Data":"877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937"} Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.731088 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" event={"ID":"0f520e1c-bdfd-4fc9-9608-237d44a73622","Type":"ContainerDied","Data":"871f87cc86619393b4f8c2501f9744ca3c69272b990f0c4ae419fe0df291d459"} Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.731127 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-mlclv" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.731138 4945 scope.go:117] "RemoveContainer" containerID="877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.759934 4945 scope.go:117] "RemoveContainer" containerID="50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.788541 4945 scope.go:117] "RemoveContainer" containerID="877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937" Dec 06 07:16:21 crc kubenswrapper[4945]: E1206 07:16:21.789129 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937\": container with ID starting with 877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937 not found: ID does not exist" containerID="877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.789174 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937"} err="failed to get container status \"877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937\": rpc error: code = NotFound desc = could not find container \"877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937\": container with ID starting with 877e1381f34384b00522e7ffcff304b5d916658dd40b561f8ebd3ac8c528d937 not found: ID does not exist" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.789202 4945 scope.go:117] "RemoveContainer" containerID="50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6" Dec 06 07:16:21 crc kubenswrapper[4945]: E1206 07:16:21.789574 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6\": container with ID starting with 50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6 not found: ID does not exist" containerID="50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.789592 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6"} err="failed to get container status \"50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6\": rpc error: code = NotFound desc = could not find container \"50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6\": container with ID starting with 50a9f9b8fa3fb66c068d442353abf27352d2207ade27aaa6c8441ca2097148f6 not found: ID does not exist" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.869309 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-dns-svc\") pod \"0f520e1c-bdfd-4fc9-9608-237d44a73622\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.869427 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-sb\") pod \"0f520e1c-bdfd-4fc9-9608-237d44a73622\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.869476 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-nb\") pod \"0f520e1c-bdfd-4fc9-9608-237d44a73622\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.869541 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-config\") pod \"0f520e1c-bdfd-4fc9-9608-237d44a73622\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.869711 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwbtj\" (UniqueName: \"kubernetes.io/projected/0f520e1c-bdfd-4fc9-9608-237d44a73622-kube-api-access-bwbtj\") pod \"0f520e1c-bdfd-4fc9-9608-237d44a73622\" (UID: \"0f520e1c-bdfd-4fc9-9608-237d44a73622\") " Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.875858 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f520e1c-bdfd-4fc9-9608-237d44a73622-kube-api-access-bwbtj" (OuterVolumeSpecName: "kube-api-access-bwbtj") pod "0f520e1c-bdfd-4fc9-9608-237d44a73622" (UID: "0f520e1c-bdfd-4fc9-9608-237d44a73622"). InnerVolumeSpecName "kube-api-access-bwbtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.950943 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-config" (OuterVolumeSpecName: "config") pod "0f520e1c-bdfd-4fc9-9608-237d44a73622" (UID: "0f520e1c-bdfd-4fc9-9608-237d44a73622"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.956264 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0f520e1c-bdfd-4fc9-9608-237d44a73622" (UID: "0f520e1c-bdfd-4fc9-9608-237d44a73622"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.956724 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0f520e1c-bdfd-4fc9-9608-237d44a73622" (UID: "0f520e1c-bdfd-4fc9-9608-237d44a73622"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.977330 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwbtj\" (UniqueName: \"kubernetes.io/projected/0f520e1c-bdfd-4fc9-9608-237d44a73622-kube-api-access-bwbtj\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.977361 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.977371 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.977379 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.985018 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0f520e1c-bdfd-4fc9-9608-237d44a73622" (UID: "0f520e1c-bdfd-4fc9-9608-237d44a73622"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.993368 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-52tjz"] Dec 06 07:16:21 crc kubenswrapper[4945]: E1206 07:16:21.993843 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f520e1c-bdfd-4fc9-9608-237d44a73622" containerName="init" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.993867 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f520e1c-bdfd-4fc9-9608-237d44a73622" containerName="init" Dec 06 07:16:21 crc kubenswrapper[4945]: E1206 07:16:21.993885 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a5cd6b1-5645-4e0a-be9d-aed6f7930c53" containerName="keystone-db-sync" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.993892 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a5cd6b1-5645-4e0a-be9d-aed6f7930c53" containerName="keystone-db-sync" Dec 06 07:16:21 crc kubenswrapper[4945]: E1206 07:16:21.993915 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f520e1c-bdfd-4fc9-9608-237d44a73622" containerName="dnsmasq-dns" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.993922 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f520e1c-bdfd-4fc9-9608-237d44a73622" containerName="dnsmasq-dns" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.994094 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a5cd6b1-5645-4e0a-be9d-aed6f7930c53" containerName="keystone-db-sync" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.994121 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f520e1c-bdfd-4fc9-9608-237d44a73622" containerName="dnsmasq-dns" Dec 06 07:16:21 crc kubenswrapper[4945]: I1206 07:16:21.995398 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.019794 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-52tjz"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.046174 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-82pl2"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.064941 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.078025 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.078272 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.078382 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-swift-storage-0\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.078427 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-sb\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.078459 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.078460 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx5hz\" (UniqueName: \"kubernetes.io/projected/2fffd627-41dc-460a-aad8-bfaeb9ec040a-kube-api-access-tx5hz\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.078567 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.078596 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wf8d7" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.078646 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-svc\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.078747 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-nb\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.079085 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-config\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.079320 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f520e1c-bdfd-4fc9-9608-237d44a73622-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.095505 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-82pl2"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.141863 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-mlclv"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.167497 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-mlclv"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182393 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-svc\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182463 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-nb\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182497 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q45tx\" (UniqueName: \"kubernetes.io/projected/5a3da785-48b2-409a-a03a-dcfd4eabe521-kube-api-access-q45tx\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182533 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-config-data\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182553 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-credential-keys\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182598 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-scripts\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182622 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-config\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182658 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-combined-ca-bundle\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182702 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-swift-storage-0\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182728 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-fernet-keys\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182754 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-sb\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.182780 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx5hz\" (UniqueName: \"kubernetes.io/projected/2fffd627-41dc-460a-aad8-bfaeb9ec040a-kube-api-access-tx5hz\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.183972 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-svc\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.184683 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-nb\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.191173 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-config\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.191300 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-swift-storage-0\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.192437 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-sb\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.218242 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-hr6hb"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.224303 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.231204 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.231273 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-x9gc5" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.236448 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.237689 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx5hz\" (UniqueName: \"kubernetes.io/projected/2fffd627-41dc-460a-aad8-bfaeb9ec040a-kube-api-access-tx5hz\") pod \"dnsmasq-dns-78b9b4595f-52tjz\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.241018 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hr6hb"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.285936 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-scripts\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.285997 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-combined-ca-bundle\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.286055 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-fernet-keys\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.286221 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q45tx\" (UniqueName: \"kubernetes.io/projected/5a3da785-48b2-409a-a03a-dcfd4eabe521-kube-api-access-q45tx\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.286260 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-config-data\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.286312 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-credential-keys\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.308308 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-combined-ca-bundle\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.309537 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-scripts\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.311588 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-mg8gj"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.319081 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-fernet-keys\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.321725 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-credential-keys\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.323222 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-config-data\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.327172 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.332918 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.333096 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-hr64m" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.333266 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.361412 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q45tx\" (UniqueName: \"kubernetes.io/projected/5a3da785-48b2-409a-a03a-dcfd4eabe521-kube-api-access-q45tx\") pod \"keystone-bootstrap-82pl2\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.379542 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.380424 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mg8gj"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.388750 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-combined-ca-bundle\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.388864 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aa503636-fff0-4195-9364-91c2f03e5c9b-etc-machine-id\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.388892 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-config-data\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.389077 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hphb\" (UniqueName: \"kubernetes.io/projected/aa503636-fff0-4195-9364-91c2f03e5c9b-kube-api-access-2hphb\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.389137 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-scripts\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.389264 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-db-sync-config-data\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.409720 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.449730 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.451774 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.455594 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.455792 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.511809 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-cks9w"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.513614 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hphb\" (UniqueName: \"kubernetes.io/projected/aa503636-fff0-4195-9364-91c2f03e5c9b-kube-api-access-2hphb\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.513672 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-scripts\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.514064 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-db-sync-config-data\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.514171 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-combined-ca-bundle\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.514328 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-config\") pod \"neutron-db-sync-mg8gj\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.514366 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aa503636-fff0-4195-9364-91c2f03e5c9b-etc-machine-id\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.514590 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-config-data\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.514986 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-combined-ca-bundle\") pod \"neutron-db-sync-mg8gj\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.515093 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spbbv\" (UniqueName: \"kubernetes.io/projected/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-kube-api-access-spbbv\") pod \"neutron-db-sync-mg8gj\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.518061 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aa503636-fff0-4195-9364-91c2f03e5c9b-etc-machine-id\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.518083 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.529681 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.536630 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-scripts\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.539545 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kgdxv" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.560302 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-db-sync-config-data\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.562219 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-config-data\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.584662 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-combined-ca-bundle\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.586883 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hphb\" (UniqueName: \"kubernetes.io/projected/aa503636-fff0-4195-9364-91c2f03e5c9b-kube-api-access-2hphb\") pod \"cinder-db-sync-hr6hb\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.604100 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619449 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sflxq\" (UniqueName: \"kubernetes.io/projected/af445462-e7c2-481c-a777-85e8566d7b71-kube-api-access-sflxq\") pod \"barbican-db-sync-cks9w\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619513 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-scripts\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619550 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spbbv\" (UniqueName: \"kubernetes.io/projected/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-kube-api-access-spbbv\") pod \"neutron-db-sync-mg8gj\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619587 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-config-data\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619611 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-db-sync-config-data\") pod \"barbican-db-sync-cks9w\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619649 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-log-httpd\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619689 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-combined-ca-bundle\") pod \"barbican-db-sync-cks9w\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619735 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n8mq\" (UniqueName: \"kubernetes.io/projected/4e525fb3-9948-4038-afce-b863a4ca5206-kube-api-access-6n8mq\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619779 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-config\") pod \"neutron-db-sync-mg8gj\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619807 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619847 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619900 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-combined-ca-bundle\") pod \"neutron-db-sync-mg8gj\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.619925 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-run-httpd\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.626148 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-combined-ca-bundle\") pod \"neutron-db-sync-mg8gj\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.637205 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-config\") pod \"neutron-db-sync-mg8gj\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.654051 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spbbv\" (UniqueName: \"kubernetes.io/projected/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-kube-api-access-spbbv\") pod \"neutron-db-sync-mg8gj\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.674514 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-cks9w"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.678176 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-5m7z8"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.679596 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.683171 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.683357 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-srcz5" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.683487 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.694349 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5m7z8"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.702064 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-52tjz"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.713900 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-nvk8s"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.715726 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721114 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n8mq\" (UniqueName: \"kubernetes.io/projected/4e525fb3-9948-4038-afce-b863a4ca5206-kube-api-access-6n8mq\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721158 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721250 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721305 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-run-httpd\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721335 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sflxq\" (UniqueName: \"kubernetes.io/projected/af445462-e7c2-481c-a777-85e8566d7b71-kube-api-access-sflxq\") pod \"barbican-db-sync-cks9w\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721355 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-scripts\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721382 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-config-data\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721400 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-db-sync-config-data\") pod \"barbican-db-sync-cks9w\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721403 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-nvk8s"] Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721430 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-log-httpd\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721584 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-combined-ca-bundle\") pod \"barbican-db-sync-cks9w\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.721850 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-log-httpd\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.726858 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.727546 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-combined-ca-bundle\") pod \"barbican-db-sync-cks9w\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.728560 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-run-httpd\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.731483 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.732674 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.735090 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-db-sync-config-data\") pod \"barbican-db-sync-cks9w\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.742897 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-scripts\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.748131 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n8mq\" (UniqueName: \"kubernetes.io/projected/4e525fb3-9948-4038-afce-b863a4ca5206-kube-api-access-6n8mq\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.749308 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.761350 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-config-data\") pod \"ceilometer-0\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.770081 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sflxq\" (UniqueName: \"kubernetes.io/projected/af445462-e7c2-481c-a777-85e8566d7b71-kube-api-access-sflxq\") pod \"barbican-db-sync-cks9w\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.781220 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.823779 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-swift-storage-0\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.824084 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-combined-ca-bundle\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.824107 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-config\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.824178 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-nb\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.824228 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx8k8\" (UniqueName: \"kubernetes.io/projected/53b7b86a-1495-4527-ad40-51ae80397771-kube-api-access-dx8k8\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.824255 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-scripts\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.824271 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-config-data\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.824307 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b7b86a-1495-4527-ad40-51ae80397771-logs\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.824336 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxcts\" (UniqueName: \"kubernetes.io/projected/a14a72cc-1734-4827-8dec-71fd48f759fa-kube-api-access-bxcts\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.824361 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-sb\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.824389 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-svc\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.896109 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cks9w" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926167 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-nb\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926223 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx8k8\" (UniqueName: \"kubernetes.io/projected/53b7b86a-1495-4527-ad40-51ae80397771-kube-api-access-dx8k8\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926253 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-scripts\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926287 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-config-data\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926312 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b7b86a-1495-4527-ad40-51ae80397771-logs\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926349 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxcts\" (UniqueName: \"kubernetes.io/projected/a14a72cc-1734-4827-8dec-71fd48f759fa-kube-api-access-bxcts\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926375 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-sb\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926403 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-svc\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926453 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-swift-storage-0\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926475 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-combined-ca-bundle\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.926491 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-config\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.927625 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-config\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.929972 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b7b86a-1495-4527-ad40-51ae80397771-logs\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.930965 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-nb\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.932686 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-sb\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.934211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-swift-storage-0\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.948236 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-svc\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.953240 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-config-data\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.964816 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-scripts\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.967075 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-combined-ca-bundle\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:22 crc kubenswrapper[4945]: I1206 07:16:22.977050 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx8k8\" (UniqueName: \"kubernetes.io/projected/53b7b86a-1495-4527-ad40-51ae80397771-kube-api-access-dx8k8\") pod \"placement-db-sync-5m7z8\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:22.999226 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f520e1c-bdfd-4fc9-9608-237d44a73622" path="/var/lib/kubelet/pods/0f520e1c-bdfd-4fc9-9608-237d44a73622/volumes" Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.019313 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxcts\" (UniqueName: \"kubernetes.io/projected/a14a72cc-1734-4827-8dec-71fd48f759fa-kube-api-access-bxcts\") pod \"dnsmasq-dns-794c6877f7-nvk8s\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.082238 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.082940 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.189589 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-52tjz"] Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.362064 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-82pl2"] Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.767048 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-82pl2" event={"ID":"5a3da785-48b2-409a-a03a-dcfd4eabe521","Type":"ContainerStarted","Data":"b56304c17f8860a7fb9704af82c3d821b0358e7f423dc2d3697c62534d14a23a"} Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.767308 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-82pl2" event={"ID":"5a3da785-48b2-409a-a03a-dcfd4eabe521","Type":"ContainerStarted","Data":"f2108aca066bf5af0e32751785e3403980f9a6031f5216dc4064fff2be3520c5"} Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.776487 4945 generic.go:334] "Generic (PLEG): container finished" podID="2fffd627-41dc-460a-aad8-bfaeb9ec040a" containerID="b57d7b2208055bf1beaeb2c74e77a7b9df3639a37f85e8f1b07d90d152168c4d" exitCode=0 Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.776537 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" event={"ID":"2fffd627-41dc-460a-aad8-bfaeb9ec040a","Type":"ContainerDied","Data":"b57d7b2208055bf1beaeb2c74e77a7b9df3639a37f85e8f1b07d90d152168c4d"} Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.776571 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" event={"ID":"2fffd627-41dc-460a-aad8-bfaeb9ec040a","Type":"ContainerStarted","Data":"c16b6580df07acb3eff6015b3c018fc8d382797c49d1c25f283f4da06c12961e"} Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.807669 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-82pl2" podStartSLOduration=1.807642213 podStartE2EDuration="1.807642213s" podCreationTimestamp="2025-12-06 07:16:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:16:23.795733849 +0000 UTC m=+1397.250594913" watchObservedRunningTime="2025-12-06 07:16:23.807642213 +0000 UTC m=+1397.262503257" Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.859378 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:16:23 crc kubenswrapper[4945]: I1206 07:16:23.869490 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mg8gj"] Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.045085 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hr6hb"] Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.075410 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5m7z8"] Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.086115 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-cks9w"] Dec 06 07:16:24 crc kubenswrapper[4945]: W1206 07:16:24.089383 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa503636_fff0_4195_9364_91c2f03e5c9b.slice/crio-3ea3a894b50806fa707f0c629148a4127b12048b065b49fc31203eb5f82d2b6f WatchSource:0}: Error finding container 3ea3a894b50806fa707f0c629148a4127b12048b065b49fc31203eb5f82d2b6f: Status 404 returned error can't find the container with id 3ea3a894b50806fa707f0c629148a4127b12048b065b49fc31203eb5f82d2b6f Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.229368 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-nvk8s"] Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.260839 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.385926 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-svc\") pod \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.386006 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx5hz\" (UniqueName: \"kubernetes.io/projected/2fffd627-41dc-460a-aad8-bfaeb9ec040a-kube-api-access-tx5hz\") pod \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.386086 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-swift-storage-0\") pod \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.386164 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-sb\") pod \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.386184 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-nb\") pod \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.386231 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-config\") pod \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\" (UID: \"2fffd627-41dc-460a-aad8-bfaeb9ec040a\") " Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.392772 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fffd627-41dc-460a-aad8-bfaeb9ec040a-kube-api-access-tx5hz" (OuterVolumeSpecName: "kube-api-access-tx5hz") pod "2fffd627-41dc-460a-aad8-bfaeb9ec040a" (UID: "2fffd627-41dc-460a-aad8-bfaeb9ec040a"). InnerVolumeSpecName "kube-api-access-tx5hz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.413110 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2fffd627-41dc-460a-aad8-bfaeb9ec040a" (UID: "2fffd627-41dc-460a-aad8-bfaeb9ec040a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.413161 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2fffd627-41dc-460a-aad8-bfaeb9ec040a" (UID: "2fffd627-41dc-460a-aad8-bfaeb9ec040a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.419057 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-config" (OuterVolumeSpecName: "config") pod "2fffd627-41dc-460a-aad8-bfaeb9ec040a" (UID: "2fffd627-41dc-460a-aad8-bfaeb9ec040a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.419762 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2fffd627-41dc-460a-aad8-bfaeb9ec040a" (UID: "2fffd627-41dc-460a-aad8-bfaeb9ec040a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.441954 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2fffd627-41dc-460a-aad8-bfaeb9ec040a" (UID: "2fffd627-41dc-460a-aad8-bfaeb9ec040a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.489588 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.489623 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.489637 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx5hz\" (UniqueName: \"kubernetes.io/projected/2fffd627-41dc-460a-aad8-bfaeb9ec040a-kube-api-access-tx5hz\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.489652 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.489664 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.489676 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fffd627-41dc-460a-aad8-bfaeb9ec040a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.802763 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" event={"ID":"2fffd627-41dc-460a-aad8-bfaeb9ec040a","Type":"ContainerDied","Data":"c16b6580df07acb3eff6015b3c018fc8d382797c49d1c25f283f4da06c12961e"} Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.802832 4945 scope.go:117] "RemoveContainer" containerID="b57d7b2208055bf1beaeb2c74e77a7b9df3639a37f85e8f1b07d90d152168c4d" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.803025 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b9b4595f-52tjz" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.813931 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e525fb3-9948-4038-afce-b863a4ca5206","Type":"ContainerStarted","Data":"9475d101b613b59f4757a824d9dc3749b65b5b2ae7a9fa19e7510f7c41bc2f45"} Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.818817 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5m7z8" event={"ID":"53b7b86a-1495-4527-ad40-51ae80397771","Type":"ContainerStarted","Data":"554cafded655c7339bc2d0c20dbe593df21490e672e6b75a4bba704392676068"} Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.828658 4945 generic.go:334] "Generic (PLEG): container finished" podID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerID="a403fc8ea55bfb2d687d2fcba3c0e339f942034f045156e2d2f7641955221a11" exitCode=0 Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.828755 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" event={"ID":"a14a72cc-1734-4827-8dec-71fd48f759fa","Type":"ContainerDied","Data":"a403fc8ea55bfb2d687d2fcba3c0e339f942034f045156e2d2f7641955221a11"} Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.828791 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" event={"ID":"a14a72cc-1734-4827-8dec-71fd48f759fa","Type":"ContainerStarted","Data":"f00942758021648652e401da69ab9d7c2e4dd7dc16b534182ff01737055032f3"} Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.833357 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hr6hb" event={"ID":"aa503636-fff0-4195-9364-91c2f03e5c9b","Type":"ContainerStarted","Data":"3ea3a894b50806fa707f0c629148a4127b12048b065b49fc31203eb5f82d2b6f"} Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.841266 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cks9w" event={"ID":"af445462-e7c2-481c-a777-85e8566d7b71","Type":"ContainerStarted","Data":"0042159a2556bbacefdcefda2eda8a4872bb974b7dc33ad94ea7f6101570355b"} Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.846613 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mg8gj" event={"ID":"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90","Type":"ContainerStarted","Data":"a55d092cd1e6de0a0878b76ed91d8a257d77965c701ebff42dc572a4081e9260"} Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.846644 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mg8gj" event={"ID":"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90","Type":"ContainerStarted","Data":"246956b2204c82ecd9802ce310b2b6588382a28e8481b67d76b3a0cf9d0a1eb9"} Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.921757 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-52tjz"] Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.961311 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-mg8gj" podStartSLOduration=2.961293583 podStartE2EDuration="2.961293583s" podCreationTimestamp="2025-12-06 07:16:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:16:24.946091049 +0000 UTC m=+1398.400952083" watchObservedRunningTime="2025-12-06 07:16:24.961293583 +0000 UTC m=+1398.416154627" Dec 06 07:16:24 crc kubenswrapper[4945]: I1206 07:16:24.980108 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-52tjz"] Dec 06 07:16:25 crc kubenswrapper[4945]: I1206 07:16:25.870945 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" event={"ID":"a14a72cc-1734-4827-8dec-71fd48f759fa","Type":"ContainerStarted","Data":"5c3fb639c9ab328189d0c15dfec716f825d76435726fdf496bb781f05072b7b0"} Dec 06 07:16:25 crc kubenswrapper[4945]: I1206 07:16:25.871491 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:25 crc kubenswrapper[4945]: I1206 07:16:25.938376 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" podStartSLOduration=3.938356818 podStartE2EDuration="3.938356818s" podCreationTimestamp="2025-12-06 07:16:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:16:25.935406758 +0000 UTC m=+1399.390267812" watchObservedRunningTime="2025-12-06 07:16:25.938356818 +0000 UTC m=+1399.393217872" Dec 06 07:16:26 crc kubenswrapper[4945]: I1206 07:16:26.195319 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:16:26 crc kubenswrapper[4945]: I1206 07:16:26.988021 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fffd627-41dc-460a-aad8-bfaeb9ec040a" path="/var/lib/kubelet/pods/2fffd627-41dc-460a-aad8-bfaeb9ec040a/volumes" Dec 06 07:16:27 crc kubenswrapper[4945]: I1206 07:16:27.935328 4945 generic.go:334] "Generic (PLEG): container finished" podID="abd7c8df-1c06-4098-acc5-8ac65805e879" containerID="948373ff86352f7383fc745821f8ce1cfcddc2703574fde9f0f77d1c9931317f" exitCode=0 Dec 06 07:16:27 crc kubenswrapper[4945]: I1206 07:16:27.935327 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g5p9c" event={"ID":"abd7c8df-1c06-4098-acc5-8ac65805e879","Type":"ContainerDied","Data":"948373ff86352f7383fc745821f8ce1cfcddc2703574fde9f0f77d1c9931317f"} Dec 06 07:16:28 crc kubenswrapper[4945]: I1206 07:16:28.951175 4945 generic.go:334] "Generic (PLEG): container finished" podID="5a3da785-48b2-409a-a03a-dcfd4eabe521" containerID="b56304c17f8860a7fb9704af82c3d821b0358e7f423dc2d3697c62534d14a23a" exitCode=0 Dec 06 07:16:28 crc kubenswrapper[4945]: I1206 07:16:28.951271 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-82pl2" event={"ID":"5a3da785-48b2-409a-a03a-dcfd4eabe521","Type":"ContainerDied","Data":"b56304c17f8860a7fb9704af82c3d821b0358e7f423dc2d3697c62534d14a23a"} Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.148601 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g5p9c" Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.328961 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-db-sync-config-data\") pod \"abd7c8df-1c06-4098-acc5-8ac65805e879\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.329066 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-config-data\") pod \"abd7c8df-1c06-4098-acc5-8ac65805e879\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.329145 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-combined-ca-bundle\") pod \"abd7c8df-1c06-4098-acc5-8ac65805e879\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.329219 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd49k\" (UniqueName: \"kubernetes.io/projected/abd7c8df-1c06-4098-acc5-8ac65805e879-kube-api-access-nd49k\") pod \"abd7c8df-1c06-4098-acc5-8ac65805e879\" (UID: \"abd7c8df-1c06-4098-acc5-8ac65805e879\") " Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.341511 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "abd7c8df-1c06-4098-acc5-8ac65805e879" (UID: "abd7c8df-1c06-4098-acc5-8ac65805e879"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.341596 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abd7c8df-1c06-4098-acc5-8ac65805e879-kube-api-access-nd49k" (OuterVolumeSpecName: "kube-api-access-nd49k") pod "abd7c8df-1c06-4098-acc5-8ac65805e879" (UID: "abd7c8df-1c06-4098-acc5-8ac65805e879"). InnerVolumeSpecName "kube-api-access-nd49k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.367438 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abd7c8df-1c06-4098-acc5-8ac65805e879" (UID: "abd7c8df-1c06-4098-acc5-8ac65805e879"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.387800 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-config-data" (OuterVolumeSpecName: "config-data") pod "abd7c8df-1c06-4098-acc5-8ac65805e879" (UID: "abd7c8df-1c06-4098-acc5-8ac65805e879"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.432522 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.432577 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.432587 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd7c8df-1c06-4098-acc5-8ac65805e879-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.432601 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd49k\" (UniqueName: \"kubernetes.io/projected/abd7c8df-1c06-4098-acc5-8ac65805e879-kube-api-access-nd49k\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.978584 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g5p9c" Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.978478 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g5p9c" event={"ID":"abd7c8df-1c06-4098-acc5-8ac65805e879","Type":"ContainerDied","Data":"cf7ba877646907a5fe68c0d03c047254fd8cf838fb9160cdeec1a92294d6d589"} Dec 06 07:16:30 crc kubenswrapper[4945]: I1206 07:16:30.978661 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf7ba877646907a5fe68c0d03c047254fd8cf838fb9160cdeec1a92294d6d589" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.783863 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-nvk8s"] Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.784326 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" podUID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerName="dnsmasq-dns" containerID="cri-o://5c3fb639c9ab328189d0c15dfec716f825d76435726fdf496bb781f05072b7b0" gracePeriod=10 Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.797482 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.884419 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-khdrt"] Dec 06 07:16:31 crc kubenswrapper[4945]: E1206 07:16:31.884983 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abd7c8df-1c06-4098-acc5-8ac65805e879" containerName="glance-db-sync" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.884997 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="abd7c8df-1c06-4098-acc5-8ac65805e879" containerName="glance-db-sync" Dec 06 07:16:31 crc kubenswrapper[4945]: E1206 07:16:31.885029 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fffd627-41dc-460a-aad8-bfaeb9ec040a" containerName="init" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.885035 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fffd627-41dc-460a-aad8-bfaeb9ec040a" containerName="init" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.885239 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fffd627-41dc-460a-aad8-bfaeb9ec040a" containerName="init" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.885254 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="abd7c8df-1c06-4098-acc5-8ac65805e879" containerName="glance-db-sync" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.886345 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.917420 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-khdrt"] Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.980696 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-config\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.980757 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8rj5\" (UniqueName: \"kubernetes.io/projected/6cdd3739-5157-4344-acfc-8ff075bc91b8-kube-api-access-z8rj5\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.980836 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-swift-storage-0\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.980862 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-nb\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.980909 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-sb\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:31 crc kubenswrapper[4945]: I1206 07:16:31.980946 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-svc\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.084313 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-sb\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.084664 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-svc\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.084724 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-config\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.084771 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8rj5\" (UniqueName: \"kubernetes.io/projected/6cdd3739-5157-4344-acfc-8ff075bc91b8-kube-api-access-z8rj5\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.084867 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-swift-storage-0\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.084890 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-nb\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.085587 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-sb\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.085672 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-nb\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.086760 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-config\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.087171 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-svc\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.087470 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-swift-storage-0\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.115394 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8rj5\" (UniqueName: \"kubernetes.io/projected/6cdd3739-5157-4344-acfc-8ff075bc91b8-kube-api-access-z8rj5\") pod \"dnsmasq-dns-7884648fd9-khdrt\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.254848 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.937208 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.977561 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.981954 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.982184 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 07:16:32 crc kubenswrapper[4945]: I1206 07:16:32.982372 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5m642" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.009833 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.084561 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" podUID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: connect: connection refused" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.107714 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.107780 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.107809 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-logs\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.107854 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.107875 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzpgr\" (UniqueName: \"kubernetes.io/projected/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-kube-api-access-gzpgr\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.107916 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.107957 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.172012 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.173671 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.176879 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.210877 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.211097 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.211184 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.211254 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-logs\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.211331 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.211386 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzpgr\" (UniqueName: \"kubernetes.io/projected/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-kube-api-access-gzpgr\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.211433 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.213269 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.213468 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.214395 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-logs\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.218229 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.218416 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.218675 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.231246 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.238670 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzpgr\" (UniqueName: \"kubernetes.io/projected/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-kube-api-access-gzpgr\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.247438 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.299557 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.312884 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.312956 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-logs\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.313055 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.313104 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvvhk\" (UniqueName: \"kubernetes.io/projected/5291f134-8e5f-4b93-b5bc-b2374117f502-kube-api-access-dvvhk\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.313124 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.313236 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.313417 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.415032 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.415084 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvvhk\" (UniqueName: \"kubernetes.io/projected/5291f134-8e5f-4b93-b5bc-b2374117f502-kube-api-access-dvvhk\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.415107 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.415149 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.415176 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.415242 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.415269 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-logs\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.415968 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-logs\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.416085 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.421481 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.421881 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.428636 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.447683 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.448105 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvvhk\" (UniqueName: \"kubernetes.io/projected/5291f134-8e5f-4b93-b5bc-b2374117f502-kube-api-access-dvvhk\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.465523 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:33 crc kubenswrapper[4945]: I1206 07:16:33.506150 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:16:34 crc kubenswrapper[4945]: I1206 07:16:34.490513 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:16:34 crc kubenswrapper[4945]: I1206 07:16:34.611601 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:16:35 crc kubenswrapper[4945]: I1206 07:16:35.092457 4945 generic.go:334] "Generic (PLEG): container finished" podID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerID="5c3fb639c9ab328189d0c15dfec716f825d76435726fdf496bb781f05072b7b0" exitCode=0 Dec 06 07:16:35 crc kubenswrapper[4945]: I1206 07:16:35.092507 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" event={"ID":"a14a72cc-1734-4827-8dec-71fd48f759fa","Type":"ContainerDied","Data":"5c3fb639c9ab328189d0c15dfec716f825d76435726fdf496bb781f05072b7b0"} Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.422579 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.515175 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-combined-ca-bundle\") pod \"5a3da785-48b2-409a-a03a-dcfd4eabe521\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.515338 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-scripts\") pod \"5a3da785-48b2-409a-a03a-dcfd4eabe521\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.515381 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-config-data\") pod \"5a3da785-48b2-409a-a03a-dcfd4eabe521\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.515452 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q45tx\" (UniqueName: \"kubernetes.io/projected/5a3da785-48b2-409a-a03a-dcfd4eabe521-kube-api-access-q45tx\") pod \"5a3da785-48b2-409a-a03a-dcfd4eabe521\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.515529 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-fernet-keys\") pod \"5a3da785-48b2-409a-a03a-dcfd4eabe521\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.515580 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-credential-keys\") pod \"5a3da785-48b2-409a-a03a-dcfd4eabe521\" (UID: \"5a3da785-48b2-409a-a03a-dcfd4eabe521\") " Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.523099 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5a3da785-48b2-409a-a03a-dcfd4eabe521" (UID: "5a3da785-48b2-409a-a03a-dcfd4eabe521"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.544831 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a3da785-48b2-409a-a03a-dcfd4eabe521-kube-api-access-q45tx" (OuterVolumeSpecName: "kube-api-access-q45tx") pod "5a3da785-48b2-409a-a03a-dcfd4eabe521" (UID: "5a3da785-48b2-409a-a03a-dcfd4eabe521"). InnerVolumeSpecName "kube-api-access-q45tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.547747 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-scripts" (OuterVolumeSpecName: "scripts") pod "5a3da785-48b2-409a-a03a-dcfd4eabe521" (UID: "5a3da785-48b2-409a-a03a-dcfd4eabe521"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.552110 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5a3da785-48b2-409a-a03a-dcfd4eabe521" (UID: "5a3da785-48b2-409a-a03a-dcfd4eabe521"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.557211 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a3da785-48b2-409a-a03a-dcfd4eabe521" (UID: "5a3da785-48b2-409a-a03a-dcfd4eabe521"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.580994 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-config-data" (OuterVolumeSpecName: "config-data") pod "5a3da785-48b2-409a-a03a-dcfd4eabe521" (UID: "5a3da785-48b2-409a-a03a-dcfd4eabe521"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.620327 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.620434 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.620452 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q45tx\" (UniqueName: \"kubernetes.io/projected/5a3da785-48b2-409a-a03a-dcfd4eabe521-kube-api-access-q45tx\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.620469 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.620481 4945 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:37 crc kubenswrapper[4945]: I1206 07:16:37.620492 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a3da785-48b2-409a-a03a-dcfd4eabe521-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:38 crc kubenswrapper[4945]: E1206 07:16:38.085587 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:82006b9c64d4c5f80483cda262d960ce6be4813665158ef1a53ea7734bbe431f" Dec 06 07:16:38 crc kubenswrapper[4945]: E1206 07:16:38.085776 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:82006b9c64d4c5f80483cda262d960ce6be4813665158ef1a53ea7734bbe431f,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sflxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-cks9w_openstack(af445462-e7c2-481c-a777-85e8566d7b71): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:16:38 crc kubenswrapper[4945]: E1206 07:16:38.086993 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-cks9w" podUID="af445462-e7c2-481c-a777-85e8566d7b71" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.137884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-82pl2" event={"ID":"5a3da785-48b2-409a-a03a-dcfd4eabe521","Type":"ContainerDied","Data":"f2108aca066bf5af0e32751785e3403980f9a6031f5216dc4064fff2be3520c5"} Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.137969 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2108aca066bf5af0e32751785e3403980f9a6031f5216dc4064fff2be3520c5" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.137979 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-82pl2" Dec 06 07:16:38 crc kubenswrapper[4945]: E1206 07:16:38.140300 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:82006b9c64d4c5f80483cda262d960ce6be4813665158ef1a53ea7734bbe431f\\\"\"" pod="openstack/barbican-db-sync-cks9w" podUID="af445462-e7c2-481c-a777-85e8566d7b71" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.520934 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-82pl2"] Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.528502 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-82pl2"] Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.618096 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2b2jk"] Dec 06 07:16:38 crc kubenswrapper[4945]: E1206 07:16:38.618593 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3da785-48b2-409a-a03a-dcfd4eabe521" containerName="keystone-bootstrap" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.618617 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3da785-48b2-409a-a03a-dcfd4eabe521" containerName="keystone-bootstrap" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.618875 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a3da785-48b2-409a-a03a-dcfd4eabe521" containerName="keystone-bootstrap" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.622718 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.625496 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.626187 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.626552 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.626740 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wf8d7" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.626935 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.649192 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2b2jk"] Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.745115 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-combined-ca-bundle\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.745201 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-scripts\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.745256 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-854xk\" (UniqueName: \"kubernetes.io/projected/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-kube-api-access-854xk\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.745320 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-config-data\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.745365 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-credential-keys\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.745468 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-fernet-keys\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.796047 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.796108 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.847164 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-fernet-keys\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.847216 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-combined-ca-bundle\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.847254 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-scripts\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.847301 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-854xk\" (UniqueName: \"kubernetes.io/projected/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-kube-api-access-854xk\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.847334 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-config-data\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.847367 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-credential-keys\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.853034 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-credential-keys\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.853138 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-fernet-keys\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.853211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-config-data\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.853576 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-scripts\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.854757 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-combined-ca-bundle\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.868118 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-854xk\" (UniqueName: \"kubernetes.io/projected/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-kube-api-access-854xk\") pod \"keystone-bootstrap-2b2jk\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.942639 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:38 crc kubenswrapper[4945]: I1206 07:16:38.965817 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a3da785-48b2-409a-a03a-dcfd4eabe521" path="/var/lib/kubelet/pods/5a3da785-48b2-409a-a03a-dcfd4eabe521/volumes" Dec 06 07:16:43 crc kubenswrapper[4945]: I1206 07:16:43.084347 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" podUID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Dec 06 07:16:47 crc kubenswrapper[4945]: E1206 07:16:47.072402 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:43a24796dabde68270dbfefa107205e173fdd6a0dc701502858cadbede69da31" Dec 06 07:16:47 crc kubenswrapper[4945]: E1206 07:16:47.072971 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:43a24796dabde68270dbfefa107205e173fdd6a0dc701502858cadbede69da31,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n58ch5c9hddh597h658h94h697h556h56dhf6hd9h5dh7fh575h558h58ch55fh5fch74hc8h555h7fh84h556h58ch668h684hf5h665h66h5f6h648q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6n8mq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(4e525fb3-9948-4038-afce-b863a4ca5206): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.202880 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.251799 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" event={"ID":"a14a72cc-1734-4827-8dec-71fd48f759fa","Type":"ContainerDied","Data":"f00942758021648652e401da69ab9d7c2e4dd7dc16b534182ff01737055032f3"} Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.251874 4945 scope.go:117] "RemoveContainer" containerID="5c3fb639c9ab328189d0c15dfec716f825d76435726fdf496bb781f05072b7b0" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.251893 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.368227 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-sb\") pod \"a14a72cc-1734-4827-8dec-71fd48f759fa\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.368308 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-nb\") pod \"a14a72cc-1734-4827-8dec-71fd48f759fa\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.368395 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-svc\") pod \"a14a72cc-1734-4827-8dec-71fd48f759fa\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.368490 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-swift-storage-0\") pod \"a14a72cc-1734-4827-8dec-71fd48f759fa\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.368518 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-config\") pod \"a14a72cc-1734-4827-8dec-71fd48f759fa\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.368679 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxcts\" (UniqueName: \"kubernetes.io/projected/a14a72cc-1734-4827-8dec-71fd48f759fa-kube-api-access-bxcts\") pod \"a14a72cc-1734-4827-8dec-71fd48f759fa\" (UID: \"a14a72cc-1734-4827-8dec-71fd48f759fa\") " Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.386895 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a14a72cc-1734-4827-8dec-71fd48f759fa-kube-api-access-bxcts" (OuterVolumeSpecName: "kube-api-access-bxcts") pod "a14a72cc-1734-4827-8dec-71fd48f759fa" (UID: "a14a72cc-1734-4827-8dec-71fd48f759fa"). InnerVolumeSpecName "kube-api-access-bxcts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.418678 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a14a72cc-1734-4827-8dec-71fd48f759fa" (UID: "a14a72cc-1734-4827-8dec-71fd48f759fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.420840 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-config" (OuterVolumeSpecName: "config") pod "a14a72cc-1734-4827-8dec-71fd48f759fa" (UID: "a14a72cc-1734-4827-8dec-71fd48f759fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.439227 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a14a72cc-1734-4827-8dec-71fd48f759fa" (UID: "a14a72cc-1734-4827-8dec-71fd48f759fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.444718 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a14a72cc-1734-4827-8dec-71fd48f759fa" (UID: "a14a72cc-1734-4827-8dec-71fd48f759fa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.449419 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a14a72cc-1734-4827-8dec-71fd48f759fa" (UID: "a14a72cc-1734-4827-8dec-71fd48f759fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.474362 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxcts\" (UniqueName: \"kubernetes.io/projected/a14a72cc-1734-4827-8dec-71fd48f759fa-kube-api-access-bxcts\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.474393 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.474403 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.474416 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.474425 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.474434 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a14a72cc-1734-4827-8dec-71fd48f759fa-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.596214 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-nvk8s"] Dec 06 07:16:47 crc kubenswrapper[4945]: I1206 07:16:47.604721 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-nvk8s"] Dec 06 07:16:48 crc kubenswrapper[4945]: I1206 07:16:48.085357 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-794c6877f7-nvk8s" podUID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Dec 06 07:16:48 crc kubenswrapper[4945]: I1206 07:16:48.965558 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a14a72cc-1734-4827-8dec-71fd48f759fa" path="/var/lib/kubelet/pods/a14a72cc-1734-4827-8dec-71fd48f759fa/volumes" Dec 06 07:16:50 crc kubenswrapper[4945]: I1206 07:16:50.104420 4945 scope.go:117] "RemoveContainer" containerID="a403fc8ea55bfb2d687d2fcba3c0e339f942034f045156e2d2f7641955221a11" Dec 06 07:16:50 crc kubenswrapper[4945]: E1206 07:16:50.286496 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2" Dec 06 07:16:50 crc kubenswrapper[4945]: E1206 07:16:50.287184 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2hphb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-hr6hb_openstack(aa503636-fff0-4195-9364-91c2f03e5c9b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:16:50 crc kubenswrapper[4945]: E1206 07:16:50.288467 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-hr6hb" podUID="aa503636-fff0-4195-9364-91c2f03e5c9b" Dec 06 07:16:50 crc kubenswrapper[4945]: I1206 07:16:50.629970 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-khdrt"] Dec 06 07:16:50 crc kubenswrapper[4945]: I1206 07:16:50.763718 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2b2jk"] Dec 06 07:16:50 crc kubenswrapper[4945]: I1206 07:16:50.873688 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:16:50 crc kubenswrapper[4945]: W1206 07:16:50.937923 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cdd3739_5157_4344_acfc_8ff075bc91b8.slice/crio-d6b1b13bc047111f12b0f297c4cf72d283334e1b18d7c887a0304a2bba3713b4 WatchSource:0}: Error finding container d6b1b13bc047111f12b0f297c4cf72d283334e1b18d7c887a0304a2bba3713b4: Status 404 returned error can't find the container with id d6b1b13bc047111f12b0f297c4cf72d283334e1b18d7c887a0304a2bba3713b4 Dec 06 07:16:50 crc kubenswrapper[4945]: W1206 07:16:50.943309 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6cd4d13_dc4a_416e_8122_4d4e8c1d5f3e.slice/crio-48e3ca51df3bcce18f633b37ed27d6e905aaf2060f9af27a6c6f061442c997b7 WatchSource:0}: Error finding container 48e3ca51df3bcce18f633b37ed27d6e905aaf2060f9af27a6c6f061442c997b7: Status 404 returned error can't find the container with id 48e3ca51df3bcce18f633b37ed27d6e905aaf2060f9af27a6c6f061442c997b7 Dec 06 07:16:51 crc kubenswrapper[4945]: I1206 07:16:51.311421 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5291f134-8e5f-4b93-b5bc-b2374117f502","Type":"ContainerStarted","Data":"bcc94168a962cadcc9a8aeaecadf49928f627ba0c7f04d2ec25dfdcdb602f0a2"} Dec 06 07:16:51 crc kubenswrapper[4945]: I1206 07:16:51.314581 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" event={"ID":"6cdd3739-5157-4344-acfc-8ff075bc91b8","Type":"ContainerStarted","Data":"2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9"} Dec 06 07:16:51 crc kubenswrapper[4945]: I1206 07:16:51.314630 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" event={"ID":"6cdd3739-5157-4344-acfc-8ff075bc91b8","Type":"ContainerStarted","Data":"d6b1b13bc047111f12b0f297c4cf72d283334e1b18d7c887a0304a2bba3713b4"} Dec 06 07:16:51 crc kubenswrapper[4945]: I1206 07:16:51.322212 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5m7z8" event={"ID":"53b7b86a-1495-4527-ad40-51ae80397771","Type":"ContainerStarted","Data":"3b3520eb06fc60e055fc656ef4c924e2b580496435b938b5e0f0c08ea8880aec"} Dec 06 07:16:51 crc kubenswrapper[4945]: I1206 07:16:51.352526 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b2jk" event={"ID":"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e","Type":"ContainerStarted","Data":"0ec1ad2614ffcda8861fab5a057374555db223dbd1a011d19ff96df9efa454d2"} Dec 06 07:16:51 crc kubenswrapper[4945]: I1206 07:16:51.352577 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b2jk" event={"ID":"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e","Type":"ContainerStarted","Data":"48e3ca51df3bcce18f633b37ed27d6e905aaf2060f9af27a6c6f061442c997b7"} Dec 06 07:16:51 crc kubenswrapper[4945]: E1206 07:16:51.360259 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2\\\"\"" pod="openstack/cinder-db-sync-hr6hb" podUID="aa503636-fff0-4195-9364-91c2f03e5c9b" Dec 06 07:16:51 crc kubenswrapper[4945]: I1206 07:16:51.401603 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2b2jk" podStartSLOduration=13.401575341 podStartE2EDuration="13.401575341s" podCreationTimestamp="2025-12-06 07:16:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:16:51.383962752 +0000 UTC m=+1424.838823816" watchObservedRunningTime="2025-12-06 07:16:51.401575341 +0000 UTC m=+1424.856436385" Dec 06 07:16:51 crc kubenswrapper[4945]: I1206 07:16:51.420575 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-5m7z8" podStartSLOduration=6.494946481 podStartE2EDuration="29.420483526s" podCreationTimestamp="2025-12-06 07:16:22 +0000 UTC" firstStartedPulling="2025-12-06 07:16:24.152934398 +0000 UTC m=+1397.607795442" lastFinishedPulling="2025-12-06 07:16:47.078471443 +0000 UTC m=+1420.533332487" observedRunningTime="2025-12-06 07:16:51.410562626 +0000 UTC m=+1424.865423660" watchObservedRunningTime="2025-12-06 07:16:51.420483526 +0000 UTC m=+1424.875344580" Dec 06 07:16:51 crc kubenswrapper[4945]: W1206 07:16:51.465896 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1c8673f_2115_496e_9ab6_b7ff4e9e74b6.slice/crio-64a1323c8bb5b921685eeebb8f36c3bb4a667ef3ded2315b3b750c2da6e37df0 WatchSource:0}: Error finding container 64a1323c8bb5b921685eeebb8f36c3bb4a667ef3ded2315b3b750c2da6e37df0: Status 404 returned error can't find the container with id 64a1323c8bb5b921685eeebb8f36c3bb4a667ef3ded2315b3b750c2da6e37df0 Dec 06 07:16:51 crc kubenswrapper[4945]: I1206 07:16:51.480614 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:16:52 crc kubenswrapper[4945]: I1206 07:16:52.364222 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6","Type":"ContainerStarted","Data":"9cc55b198ed9f0064e3a9058d01cce5bb8d525b240af76c057b68860a95b54c8"} Dec 06 07:16:52 crc kubenswrapper[4945]: I1206 07:16:52.364822 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6","Type":"ContainerStarted","Data":"64a1323c8bb5b921685eeebb8f36c3bb4a667ef3ded2315b3b750c2da6e37df0"} Dec 06 07:16:52 crc kubenswrapper[4945]: I1206 07:16:52.366810 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5291f134-8e5f-4b93-b5bc-b2374117f502","Type":"ContainerStarted","Data":"949300175b977c4e17d92491ddd4ce647b0d87ea1ca4f0ff07636fdbe707cbcc"} Dec 06 07:16:52 crc kubenswrapper[4945]: I1206 07:16:52.371479 4945 generic.go:334] "Generic (PLEG): container finished" podID="6cdd3739-5157-4344-acfc-8ff075bc91b8" containerID="2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9" exitCode=0 Dec 06 07:16:52 crc kubenswrapper[4945]: I1206 07:16:52.371552 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" event={"ID":"6cdd3739-5157-4344-acfc-8ff075bc91b8","Type":"ContainerDied","Data":"2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9"} Dec 06 07:16:52 crc kubenswrapper[4945]: I1206 07:16:52.371571 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" event={"ID":"6cdd3739-5157-4344-acfc-8ff075bc91b8","Type":"ContainerStarted","Data":"f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a"} Dec 06 07:16:52 crc kubenswrapper[4945]: I1206 07:16:52.371679 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:52 crc kubenswrapper[4945]: I1206 07:16:52.376411 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e525fb3-9948-4038-afce-b863a4ca5206","Type":"ContainerStarted","Data":"fdee576b16798f10cbf0d6df65b22dc489c5de6fce0d46d09574de5ceaf78ca7"} Dec 06 07:16:52 crc kubenswrapper[4945]: I1206 07:16:52.405387 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" podStartSLOduration=21.40525755 podStartE2EDuration="21.40525755s" podCreationTimestamp="2025-12-06 07:16:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:16:52.396943314 +0000 UTC m=+1425.851804368" watchObservedRunningTime="2025-12-06 07:16:52.40525755 +0000 UTC m=+1425.860118594" Dec 06 07:16:53 crc kubenswrapper[4945]: I1206 07:16:53.413153 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6","Type":"ContainerStarted","Data":"2efb64e9af398c6eccee02f8d25fb98cd60674bc6dfd8415c30e1b08adb56789"} Dec 06 07:16:53 crc kubenswrapper[4945]: I1206 07:16:53.413415 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" containerName="glance-log" containerID="cri-o://9cc55b198ed9f0064e3a9058d01cce5bb8d525b240af76c057b68860a95b54c8" gracePeriod=30 Dec 06 07:16:53 crc kubenswrapper[4945]: I1206 07:16:53.413709 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" containerName="glance-httpd" containerID="cri-o://2efb64e9af398c6eccee02f8d25fb98cd60674bc6dfd8415c30e1b08adb56789" gracePeriod=30 Dec 06 07:16:53 crc kubenswrapper[4945]: I1206 07:16:53.428486 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5291f134-8e5f-4b93-b5bc-b2374117f502","Type":"ContainerStarted","Data":"53c0075dccd3a254fb9efe790919572c497272e3bf93c459ad8b10ca8e292557"} Dec 06 07:16:53 crc kubenswrapper[4945]: I1206 07:16:53.428645 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5291f134-8e5f-4b93-b5bc-b2374117f502" containerName="glance-log" containerID="cri-o://949300175b977c4e17d92491ddd4ce647b0d87ea1ca4f0ff07636fdbe707cbcc" gracePeriod=30 Dec 06 07:16:53 crc kubenswrapper[4945]: I1206 07:16:53.428735 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5291f134-8e5f-4b93-b5bc-b2374117f502" containerName="glance-httpd" containerID="cri-o://53c0075dccd3a254fb9efe790919572c497272e3bf93c459ad8b10ca8e292557" gracePeriod=30 Dec 06 07:16:53 crc kubenswrapper[4945]: I1206 07:16:53.446450 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=22.446428009 podStartE2EDuration="22.446428009s" podCreationTimestamp="2025-12-06 07:16:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:16:53.441915947 +0000 UTC m=+1426.896777001" watchObservedRunningTime="2025-12-06 07:16:53.446428009 +0000 UTC m=+1426.901289053" Dec 06 07:16:53 crc kubenswrapper[4945]: I1206 07:16:53.491406 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=21.491382563 podStartE2EDuration="21.491382563s" podCreationTimestamp="2025-12-06 07:16:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:16:53.481696959 +0000 UTC m=+1426.936558033" watchObservedRunningTime="2025-12-06 07:16:53.491382563 +0000 UTC m=+1426.946243617" Dec 06 07:16:54 crc kubenswrapper[4945]: I1206 07:16:54.443126 4945 generic.go:334] "Generic (PLEG): container finished" podID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" containerID="9cc55b198ed9f0064e3a9058d01cce5bb8d525b240af76c057b68860a95b54c8" exitCode=143 Dec 06 07:16:54 crc kubenswrapper[4945]: I1206 07:16:54.443205 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6","Type":"ContainerDied","Data":"9cc55b198ed9f0064e3a9058d01cce5bb8d525b240af76c057b68860a95b54c8"} Dec 06 07:16:54 crc kubenswrapper[4945]: I1206 07:16:54.446383 4945 generic.go:334] "Generic (PLEG): container finished" podID="5291f134-8e5f-4b93-b5bc-b2374117f502" containerID="949300175b977c4e17d92491ddd4ce647b0d87ea1ca4f0ff07636fdbe707cbcc" exitCode=143 Dec 06 07:16:54 crc kubenswrapper[4945]: I1206 07:16:54.446431 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5291f134-8e5f-4b93-b5bc-b2374117f502","Type":"ContainerDied","Data":"949300175b977c4e17d92491ddd4ce647b0d87ea1ca4f0ff07636fdbe707cbcc"} Dec 06 07:16:55 crc kubenswrapper[4945]: I1206 07:16:55.459024 4945 generic.go:334] "Generic (PLEG): container finished" podID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" containerID="2efb64e9af398c6eccee02f8d25fb98cd60674bc6dfd8415c30e1b08adb56789" exitCode=0 Dec 06 07:16:55 crc kubenswrapper[4945]: I1206 07:16:55.459090 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6","Type":"ContainerDied","Data":"2efb64e9af398c6eccee02f8d25fb98cd60674bc6dfd8415c30e1b08adb56789"} Dec 06 07:16:55 crc kubenswrapper[4945]: I1206 07:16:55.462150 4945 generic.go:334] "Generic (PLEG): container finished" podID="5291f134-8e5f-4b93-b5bc-b2374117f502" containerID="53c0075dccd3a254fb9efe790919572c497272e3bf93c459ad8b10ca8e292557" exitCode=0 Dec 06 07:16:55 crc kubenswrapper[4945]: I1206 07:16:55.462194 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5291f134-8e5f-4b93-b5bc-b2374117f502","Type":"ContainerDied","Data":"53c0075dccd3a254fb9efe790919572c497272e3bf93c459ad8b10ca8e292557"} Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.436933 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.478796 4945 generic.go:334] "Generic (PLEG): container finished" podID="c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" containerID="0ec1ad2614ffcda8861fab5a057374555db223dbd1a011d19ff96df9efa454d2" exitCode=0 Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.478883 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b2jk" event={"ID":"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e","Type":"ContainerDied","Data":"0ec1ad2614ffcda8861fab5a057374555db223dbd1a011d19ff96df9efa454d2"} Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.490883 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.490871 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6","Type":"ContainerDied","Data":"64a1323c8bb5b921685eeebb8f36c3bb4a667ef3ded2315b3b750c2da6e37df0"} Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.491078 4945 scope.go:117] "RemoveContainer" containerID="2efb64e9af398c6eccee02f8d25fb98cd60674bc6dfd8415c30e1b08adb56789" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.495791 4945 generic.go:334] "Generic (PLEG): container finished" podID="53b7b86a-1495-4527-ad40-51ae80397771" containerID="3b3520eb06fc60e055fc656ef4c924e2b580496435b938b5e0f0c08ea8880aec" exitCode=0 Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.495848 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5m7z8" event={"ID":"53b7b86a-1495-4527-ad40-51ae80397771","Type":"ContainerDied","Data":"3b3520eb06fc60e055fc656ef4c924e2b580496435b938b5e0f0c08ea8880aec"} Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.592186 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-httpd-run\") pod \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.592253 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzpgr\" (UniqueName: \"kubernetes.io/projected/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-kube-api-access-gzpgr\") pod \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.592332 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.592374 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-combined-ca-bundle\") pod \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.592455 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-config-data\") pod \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.592559 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-scripts\") pod \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.592604 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-logs\") pod \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\" (UID: \"a1c8673f-2115-496e-9ab6-b7ff4e9e74b6\") " Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.592781 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" (UID: "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.593070 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-logs" (OuterVolumeSpecName: "logs") pod "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" (UID: "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.594094 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.594120 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.599038 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" (UID: "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.603591 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-kube-api-access-gzpgr" (OuterVolumeSpecName: "kube-api-access-gzpgr") pod "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" (UID: "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6"). InnerVolumeSpecName "kube-api-access-gzpgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.607921 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-scripts" (OuterVolumeSpecName: "scripts") pod "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" (UID: "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.629010 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" (UID: "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.659465 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-config-data" (OuterVolumeSpecName: "config-data") pod "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" (UID: "a1c8673f-2115-496e-9ab6-b7ff4e9e74b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.695923 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.696226 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.696241 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzpgr\" (UniqueName: \"kubernetes.io/projected/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-kube-api-access-gzpgr\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.696297 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.696312 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.716431 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.798459 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.844434 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.859594 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.880171 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:16:56 crc kubenswrapper[4945]: E1206 07:16:56.881266 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerName="dnsmasq-dns" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.881301 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerName="dnsmasq-dns" Dec 06 07:16:56 crc kubenswrapper[4945]: E1206 07:16:56.881326 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" containerName="glance-httpd" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.881336 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" containerName="glance-httpd" Dec 06 07:16:56 crc kubenswrapper[4945]: E1206 07:16:56.881355 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" containerName="glance-log" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.881363 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" containerName="glance-log" Dec 06 07:16:56 crc kubenswrapper[4945]: E1206 07:16:56.881388 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerName="init" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.881395 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerName="init" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.881619 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a14a72cc-1734-4827-8dec-71fd48f759fa" containerName="dnsmasq-dns" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.881635 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" containerName="glance-httpd" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.881649 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" containerName="glance-log" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.884569 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.889271 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.890736 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.893962 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 07:16:56 crc kubenswrapper[4945]: I1206 07:16:56.977620 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c8673f-2115-496e-9ab6-b7ff4e9e74b6" path="/var/lib/kubelet/pods/a1c8673f-2115-496e-9ab6-b7ff4e9e74b6/volumes" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.002403 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-logs\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.002457 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.002516 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k89ct\" (UniqueName: \"kubernetes.io/projected/5083dbae-2860-4efd-84e5-503b3a11d99b-kube-api-access-k89ct\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.002538 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.002651 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.002931 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.003268 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-scripts\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.003495 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-config-data\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.106965 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-config-data\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.107563 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-logs\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.107599 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.107650 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k89ct\" (UniqueName: \"kubernetes.io/projected/5083dbae-2860-4efd-84e5-503b3a11d99b-kube-api-access-k89ct\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.107694 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.107756 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.107909 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.107972 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.109571 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-scripts\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.110547 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.111051 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-logs\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.113431 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.118495 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.120594 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-scripts\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.128820 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k89ct\" (UniqueName: \"kubernetes.io/projected/5083dbae-2860-4efd-84e5-503b3a11d99b-kube-api-access-k89ct\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.130228 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-config-data\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.136119 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.205441 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.257389 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.372352 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-nhnm5"] Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.372670 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" podUID="a68a3b24-6715-478e-9b85-edcc08b905dd" containerName="dnsmasq-dns" containerID="cri-o://13de05e3688414d4b0922e8767a080a6d6be044b6a8e10519316924b09fc2d99" gracePeriod=10 Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.532008 4945 generic.go:334] "Generic (PLEG): container finished" podID="a68a3b24-6715-478e-9b85-edcc08b905dd" containerID="13de05e3688414d4b0922e8767a080a6d6be044b6a8e10519316924b09fc2d99" exitCode=0 Dec 06 07:16:57 crc kubenswrapper[4945]: I1206 07:16:57.532111 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" event={"ID":"a68a3b24-6715-478e-9b85-edcc08b905dd","Type":"ContainerDied","Data":"13de05e3688414d4b0922e8767a080a6d6be044b6a8e10519316924b09fc2d99"} Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.428153 4945 scope.go:117] "RemoveContainer" containerID="9cc55b198ed9f0064e3a9058d01cce5bb8d525b240af76c057b68860a95b54c8" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.558354 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5m7z8" event={"ID":"53b7b86a-1495-4527-ad40-51ae80397771","Type":"ContainerDied","Data":"554cafded655c7339bc2d0c20dbe593df21490e672e6b75a4bba704392676068"} Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.558400 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="554cafded655c7339bc2d0c20dbe593df21490e672e6b75a4bba704392676068" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.560201 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b2jk" event={"ID":"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e","Type":"ContainerDied","Data":"48e3ca51df3bcce18f633b37ed27d6e905aaf2060f9af27a6c6f061442c997b7"} Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.560230 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48e3ca51df3bcce18f633b37ed27d6e905aaf2060f9af27a6c6f061442c997b7" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.584498 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5291f134-8e5f-4b93-b5bc-b2374117f502","Type":"ContainerDied","Data":"bcc94168a962cadcc9a8aeaecadf49928f627ba0c7f04d2ec25dfdcdb602f0a2"} Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.584587 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcc94168a962cadcc9a8aeaecadf49928f627ba0c7f04d2ec25dfdcdb602f0a2" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.675610 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.682629 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.689214 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.745998 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-scripts\") pod \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746062 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-httpd-run\") pod \"5291f134-8e5f-4b93-b5bc-b2374117f502\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746091 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-combined-ca-bundle\") pod \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746167 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-config-data\") pod \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746191 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-854xk\" (UniqueName: \"kubernetes.io/projected/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-kube-api-access-854xk\") pod \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746213 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-scripts\") pod \"5291f134-8e5f-4b93-b5bc-b2374117f502\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746246 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvvhk\" (UniqueName: \"kubernetes.io/projected/5291f134-8e5f-4b93-b5bc-b2374117f502-kube-api-access-dvvhk\") pod \"5291f134-8e5f-4b93-b5bc-b2374117f502\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746262 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-logs\") pod \"5291f134-8e5f-4b93-b5bc-b2374117f502\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746322 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-config-data\") pod \"5291f134-8e5f-4b93-b5bc-b2374117f502\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746350 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"5291f134-8e5f-4b93-b5bc-b2374117f502\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746385 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-fernet-keys\") pod \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746436 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-credential-keys\") pod \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\" (UID: \"c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.746476 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-combined-ca-bundle\") pod \"5291f134-8e5f-4b93-b5bc-b2374117f502\" (UID: \"5291f134-8e5f-4b93-b5bc-b2374117f502\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.749569 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-logs" (OuterVolumeSpecName: "logs") pod "5291f134-8e5f-4b93-b5bc-b2374117f502" (UID: "5291f134-8e5f-4b93-b5bc-b2374117f502"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.749822 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5291f134-8e5f-4b93-b5bc-b2374117f502" (UID: "5291f134-8e5f-4b93-b5bc-b2374117f502"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.767508 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.790976 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "5291f134-8e5f-4b93-b5bc-b2374117f502" (UID: "5291f134-8e5f-4b93-b5bc-b2374117f502"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.796923 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-kube-api-access-854xk" (OuterVolumeSpecName: "kube-api-access-854xk") pod "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" (UID: "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e"). InnerVolumeSpecName "kube-api-access-854xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.797135 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-scripts" (OuterVolumeSpecName: "scripts") pod "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" (UID: "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.804649 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5291f134-8e5f-4b93-b5bc-b2374117f502-kube-api-access-dvvhk" (OuterVolumeSpecName: "kube-api-access-dvvhk") pod "5291f134-8e5f-4b93-b5bc-b2374117f502" (UID: "5291f134-8e5f-4b93-b5bc-b2374117f502"). InnerVolumeSpecName "kube-api-access-dvvhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.808388 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-scripts" (OuterVolumeSpecName: "scripts") pod "5291f134-8e5f-4b93-b5bc-b2374117f502" (UID: "5291f134-8e5f-4b93-b5bc-b2374117f502"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.808876 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" (UID: "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.812203 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" (UID: "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.848192 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc8rn\" (UniqueName: \"kubernetes.io/projected/a68a3b24-6715-478e-9b85-edcc08b905dd-kube-api-access-dc8rn\") pod \"a68a3b24-6715-478e-9b85-edcc08b905dd\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.848366 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-svc\") pod \"a68a3b24-6715-478e-9b85-edcc08b905dd\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.848439 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-config-data\") pod \"53b7b86a-1495-4527-ad40-51ae80397771\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.848510 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-nb\") pod \"a68a3b24-6715-478e-9b85-edcc08b905dd\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.848609 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-combined-ca-bundle\") pod \"53b7b86a-1495-4527-ad40-51ae80397771\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.848684 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-scripts\") pod \"53b7b86a-1495-4527-ad40-51ae80397771\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.848729 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b7b86a-1495-4527-ad40-51ae80397771-logs\") pod \"53b7b86a-1495-4527-ad40-51ae80397771\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.848845 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-swift-storage-0\") pod \"a68a3b24-6715-478e-9b85-edcc08b905dd\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.848899 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx8k8\" (UniqueName: \"kubernetes.io/projected/53b7b86a-1495-4527-ad40-51ae80397771-kube-api-access-dx8k8\") pod \"53b7b86a-1495-4527-ad40-51ae80397771\" (UID: \"53b7b86a-1495-4527-ad40-51ae80397771\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.848942 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-config\") pod \"a68a3b24-6715-478e-9b85-edcc08b905dd\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.849145 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-sb\") pod \"a68a3b24-6715-478e-9b85-edcc08b905dd\" (UID: \"a68a3b24-6715-478e-9b85-edcc08b905dd\") " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.849515 4945 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.849534 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.849544 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.849553 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-854xk\" (UniqueName: \"kubernetes.io/projected/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-kube-api-access-854xk\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.849564 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.849572 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvvhk\" (UniqueName: \"kubernetes.io/projected/5291f134-8e5f-4b93-b5bc-b2374117f502-kube-api-access-dvvhk\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.849581 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5291f134-8e5f-4b93-b5bc-b2374117f502-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.849600 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.849610 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.858528 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53b7b86a-1495-4527-ad40-51ae80397771-logs" (OuterVolumeSpecName: "logs") pod "53b7b86a-1495-4527-ad40-51ae80397771" (UID: "53b7b86a-1495-4527-ad40-51ae80397771"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.859586 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-scripts" (OuterVolumeSpecName: "scripts") pod "53b7b86a-1495-4527-ad40-51ae80397771" (UID: "53b7b86a-1495-4527-ad40-51ae80397771"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.871850 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a68a3b24-6715-478e-9b85-edcc08b905dd-kube-api-access-dc8rn" (OuterVolumeSpecName: "kube-api-access-dc8rn") pod "a68a3b24-6715-478e-9b85-edcc08b905dd" (UID: "a68a3b24-6715-478e-9b85-edcc08b905dd"). InnerVolumeSpecName "kube-api-access-dc8rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.882040 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53b7b86a-1495-4527-ad40-51ae80397771-kube-api-access-dx8k8" (OuterVolumeSpecName: "kube-api-access-dx8k8") pod "53b7b86a-1495-4527-ad40-51ae80397771" (UID: "53b7b86a-1495-4527-ad40-51ae80397771"). InnerVolumeSpecName "kube-api-access-dx8k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.931306 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.934073 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5291f134-8e5f-4b93-b5bc-b2374117f502" (UID: "5291f134-8e5f-4b93-b5bc-b2374117f502"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.935174 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" (UID: "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.940502 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-config-data" (OuterVolumeSpecName: "config-data") pod "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" (UID: "c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.951624 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.951649 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.951660 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b7b86a-1495-4527-ad40-51ae80397771-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.951670 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.951679 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx8k8\" (UniqueName: \"kubernetes.io/projected/53b7b86a-1495-4527-ad40-51ae80397771-kube-api-access-dx8k8\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.951688 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.951696 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc8rn\" (UniqueName: \"kubernetes.io/projected/a68a3b24-6715-478e-9b85-edcc08b905dd-kube-api-access-dc8rn\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.951706 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.975287 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-config-data" (OuterVolumeSpecName: "config-data") pod "53b7b86a-1495-4527-ad40-51ae80397771" (UID: "53b7b86a-1495-4527-ad40-51ae80397771"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.975263 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-config-data" (OuterVolumeSpecName: "config-data") pod "5291f134-8e5f-4b93-b5bc-b2374117f502" (UID: "5291f134-8e5f-4b93-b5bc-b2374117f502"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.982620 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53b7b86a-1495-4527-ad40-51ae80397771" (UID: "53b7b86a-1495-4527-ad40-51ae80397771"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.988222 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a68a3b24-6715-478e-9b85-edcc08b905dd" (UID: "a68a3b24-6715-478e-9b85-edcc08b905dd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:58 crc kubenswrapper[4945]: I1206 07:16:58.997489 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a68a3b24-6715-478e-9b85-edcc08b905dd" (UID: "a68a3b24-6715-478e-9b85-edcc08b905dd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.002452 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a68a3b24-6715-478e-9b85-edcc08b905dd" (UID: "a68a3b24-6715-478e-9b85-edcc08b905dd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.005010 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-config" (OuterVolumeSpecName: "config") pod "a68a3b24-6715-478e-9b85-edcc08b905dd" (UID: "a68a3b24-6715-478e-9b85-edcc08b905dd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.005993 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a68a3b24-6715-478e-9b85-edcc08b905dd" (UID: "a68a3b24-6715-478e-9b85-edcc08b905dd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.056212 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.057550 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5291f134-8e5f-4b93-b5bc-b2374117f502-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.057646 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.057718 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.057863 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.057932 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b7b86a-1495-4527-ad40-51ae80397771-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.057986 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.058271 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68a3b24-6715-478e-9b85-edcc08b905dd-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:16:59 crc kubenswrapper[4945]: W1206 07:16:59.112343 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5083dbae_2860_4efd_84e5_503b3a11d99b.slice/crio-19599e6d8c229bc6c5cbe70fd54ceb47dd313d99909ad7f2c8e98b518d9fe21d WatchSource:0}: Error finding container 19599e6d8c229bc6c5cbe70fd54ceb47dd313d99909ad7f2c8e98b518d9fe21d: Status 404 returned error can't find the container with id 19599e6d8c229bc6c5cbe70fd54ceb47dd313d99909ad7f2c8e98b518d9fe21d Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.123898 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.594408 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5083dbae-2860-4efd-84e5-503b3a11d99b","Type":"ContainerStarted","Data":"19599e6d8c229bc6c5cbe70fd54ceb47dd313d99909ad7f2c8e98b518d9fe21d"} Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.599058 4945 generic.go:334] "Generic (PLEG): container finished" podID="c24e6d0b-d002-4d0c-8f6d-bb94de58fb90" containerID="a55d092cd1e6de0a0878b76ed91d8a257d77965c701ebff42dc572a4081e9260" exitCode=0 Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.599123 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mg8gj" event={"ID":"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90","Type":"ContainerDied","Data":"a55d092cd1e6de0a0878b76ed91d8a257d77965c701ebff42dc572a4081e9260"} Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.607532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e525fb3-9948-4038-afce-b863a4ca5206","Type":"ContainerStarted","Data":"2ddc90f531ce8123229116bd612f56498d8293ed1cb1cd9acab8b626c6700c6c"} Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.610439 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" event={"ID":"a68a3b24-6715-478e-9b85-edcc08b905dd","Type":"ContainerDied","Data":"789bcedca791e94cd1bba82384dcbcaacdf868936c9a7c5cb244e268104c5557"} Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.610489 4945 scope.go:117] "RemoveContainer" containerID="13de05e3688414d4b0922e8767a080a6d6be044b6a8e10519316924b09fc2d99" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.610620 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-nhnm5" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.621157 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.622977 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b2jk" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.623620 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cks9w" event={"ID":"af445462-e7c2-481c-a777-85e8566d7b71","Type":"ContainerStarted","Data":"51c9dc30540f672f0cad19373fdd58c9f35dc611b2c4f98ab2d185612f7a7af6"} Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.624022 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5m7z8" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.660680 4945 scope.go:117] "RemoveContainer" containerID="1edbf1ee4c10c69b75fab6dbda0dd63e28ec920b8c203e295ec3bd09380500a5" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.689653 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-cks9w" podStartSLOduration=3.410735421 podStartE2EDuration="37.689629672s" podCreationTimestamp="2025-12-06 07:16:22 +0000 UTC" firstStartedPulling="2025-12-06 07:16:24.149248828 +0000 UTC m=+1397.604109872" lastFinishedPulling="2025-12-06 07:16:58.428143079 +0000 UTC m=+1431.883004123" observedRunningTime="2025-12-06 07:16:59.662483253 +0000 UTC m=+1433.117344317" watchObservedRunningTime="2025-12-06 07:16:59.689629672 +0000 UTC m=+1433.144490716" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.699789 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-nhnm5"] Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.716187 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-nhnm5"] Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.726393 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.742948 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.753586 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:16:59 crc kubenswrapper[4945]: E1206 07:16:59.754134 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" containerName="keystone-bootstrap" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754162 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" containerName="keystone-bootstrap" Dec 06 07:16:59 crc kubenswrapper[4945]: E1206 07:16:59.754189 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a68a3b24-6715-478e-9b85-edcc08b905dd" containerName="init" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754198 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a68a3b24-6715-478e-9b85-edcc08b905dd" containerName="init" Dec 06 07:16:59 crc kubenswrapper[4945]: E1206 07:16:59.754246 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5291f134-8e5f-4b93-b5bc-b2374117f502" containerName="glance-log" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754256 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5291f134-8e5f-4b93-b5bc-b2374117f502" containerName="glance-log" Dec 06 07:16:59 crc kubenswrapper[4945]: E1206 07:16:59.754267 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a68a3b24-6715-478e-9b85-edcc08b905dd" containerName="dnsmasq-dns" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754289 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a68a3b24-6715-478e-9b85-edcc08b905dd" containerName="dnsmasq-dns" Dec 06 07:16:59 crc kubenswrapper[4945]: E1206 07:16:59.754313 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5291f134-8e5f-4b93-b5bc-b2374117f502" containerName="glance-httpd" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754322 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5291f134-8e5f-4b93-b5bc-b2374117f502" containerName="glance-httpd" Dec 06 07:16:59 crc kubenswrapper[4945]: E1206 07:16:59.754342 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53b7b86a-1495-4527-ad40-51ae80397771" containerName="placement-db-sync" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754350 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="53b7b86a-1495-4527-ad40-51ae80397771" containerName="placement-db-sync" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754571 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="53b7b86a-1495-4527-ad40-51ae80397771" containerName="placement-db-sync" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754591 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5291f134-8e5f-4b93-b5bc-b2374117f502" containerName="glance-httpd" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754614 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a68a3b24-6715-478e-9b85-edcc08b905dd" containerName="dnsmasq-dns" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754623 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" containerName="keystone-bootstrap" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.754638 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5291f134-8e5f-4b93-b5bc-b2374117f502" containerName="glance-log" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.755864 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.759602 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.762423 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.767209 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.881773 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.881820 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.881843 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.881918 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-logs\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.881965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5wfl\" (UniqueName: \"kubernetes.io/projected/8809238d-f467-4a48-b009-15c8b974aed0-kube-api-access-l5wfl\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.881999 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.882086 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.882168 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.917515 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-694b7fd487-99w9l"] Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.918930 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.933550 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wf8d7" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.933790 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.943775 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.943985 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.960620 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.963345 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.963523 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-694b7fd487-99w9l"] Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.971830 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-75599744c6-dmqzn"] Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.973628 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.978926 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.980363 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-srcz5" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.994500 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.994680 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.994791 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.995622 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.995684 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-logs\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.995745 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5wfl\" (UniqueName: \"kubernetes.io/projected/8809238d-f467-4a48-b009-15c8b974aed0-kube-api-access-l5wfl\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.995791 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.995822 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.995868 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.995982 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.996017 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:16:59 crc kubenswrapper[4945]: I1206 07:16:59.997137 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.010090 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.010344 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-logs\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.021172 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.022307 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.027144 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.034846 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.037882 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75599744c6-dmqzn"] Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.066842 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5wfl\" (UniqueName: \"kubernetes.io/projected/8809238d-f467-4a48-b009-15c8b974aed0-kube-api-access-l5wfl\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098529 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-combined-ca-bundle\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098584 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-internal-tls-certs\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098613 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-internal-tls-certs\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098653 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfx87\" (UniqueName: \"kubernetes.io/projected/569f0e1c-e58f-4e87-bde3-e7be4659639a-kube-api-access-lfx87\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098678 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-public-tls-certs\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098695 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/569f0e1c-e58f-4e87-bde3-e7be4659639a-logs\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098744 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-fernet-keys\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098781 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-scripts\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098800 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-scripts\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098822 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-config-data\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098843 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-public-tls-certs\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098859 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-config-data\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098878 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpwxr\" (UniqueName: \"kubernetes.io/projected/5af85716-fd7e-45e9-8320-4db8800302a8-kube-api-access-fpwxr\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098900 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-combined-ca-bundle\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.098926 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-credential-keys\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.128086 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200452 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-fernet-keys\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200735 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-scripts\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200759 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-scripts\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200782 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-config-data\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200802 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-public-tls-certs\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200818 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-config-data\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200841 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpwxr\" (UniqueName: \"kubernetes.io/projected/5af85716-fd7e-45e9-8320-4db8800302a8-kube-api-access-fpwxr\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200862 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-combined-ca-bundle\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200885 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-credential-keys\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200924 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-combined-ca-bundle\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200954 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-internal-tls-certs\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.200973 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-internal-tls-certs\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.201004 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfx87\" (UniqueName: \"kubernetes.io/projected/569f0e1c-e58f-4e87-bde3-e7be4659639a-kube-api-access-lfx87\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.201025 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-public-tls-certs\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.201044 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/569f0e1c-e58f-4e87-bde3-e7be4659639a-logs\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.201744 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/569f0e1c-e58f-4e87-bde3-e7be4659639a-logs\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.205900 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-scripts\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.213977 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-internal-tls-certs\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.214553 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-scripts\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.215568 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-credential-keys\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.223109 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-config-data\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.223609 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-public-tls-certs\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.224748 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-fernet-keys\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.224944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-combined-ca-bundle\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.229769 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-public-tls-certs\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.245032 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-combined-ca-bundle\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.249549 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-config-data\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.250173 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-internal-tls-certs\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.253693 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpwxr\" (UniqueName: \"kubernetes.io/projected/5af85716-fd7e-45e9-8320-4db8800302a8-kube-api-access-fpwxr\") pod \"keystone-694b7fd487-99w9l\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.253782 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfx87\" (UniqueName: \"kubernetes.io/projected/569f0e1c-e58f-4e87-bde3-e7be4659639a-kube-api-access-lfx87\") pod \"placement-75599744c6-dmqzn\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.375897 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.435308 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.543243 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.641655 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5083dbae-2860-4efd-84e5-503b3a11d99b","Type":"ContainerStarted","Data":"c9f3cc29bda0b30b5a9c7145f3e4d1a71bf681a69aebb900b3ca0ec8bdb9d06b"} Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.970239 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5291f134-8e5f-4b93-b5bc-b2374117f502" path="/var/lib/kubelet/pods/5291f134-8e5f-4b93-b5bc-b2374117f502/volumes" Dec 06 07:17:00 crc kubenswrapper[4945]: I1206 07:17:00.971290 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a68a3b24-6715-478e-9b85-edcc08b905dd" path="/var/lib/kubelet/pods/a68a3b24-6715-478e-9b85-edcc08b905dd/volumes" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.001116 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.069846 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75599744c6-dmqzn"] Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.110196 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.229768 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spbbv\" (UniqueName: \"kubernetes.io/projected/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-kube-api-access-spbbv\") pod \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.229821 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-combined-ca-bundle\") pod \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.230115 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-config\") pod \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\" (UID: \"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90\") " Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.243044 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-kube-api-access-spbbv" (OuterVolumeSpecName: "kube-api-access-spbbv") pod "c24e6d0b-d002-4d0c-8f6d-bb94de58fb90" (UID: "c24e6d0b-d002-4d0c-8f6d-bb94de58fb90"). InnerVolumeSpecName "kube-api-access-spbbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.267561 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-694b7fd487-99w9l"] Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.288859 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-config" (OuterVolumeSpecName: "config") pod "c24e6d0b-d002-4d0c-8f6d-bb94de58fb90" (UID: "c24e6d0b-d002-4d0c-8f6d-bb94de58fb90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.302983 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c24e6d0b-d002-4d0c-8f6d-bb94de58fb90" (UID: "c24e6d0b-d002-4d0c-8f6d-bb94de58fb90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.336466 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.336499 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spbbv\" (UniqueName: \"kubernetes.io/projected/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-kube-api-access-spbbv\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.336877 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.672256 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mg8gj" event={"ID":"c24e6d0b-d002-4d0c-8f6d-bb94de58fb90","Type":"ContainerDied","Data":"246956b2204c82ecd9802ce310b2b6588382a28e8481b67d76b3a0cf9d0a1eb9"} Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.672619 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="246956b2204c82ecd9802ce310b2b6588382a28e8481b67d76b3a0cf9d0a1eb9" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.672932 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mg8gj" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.683734 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75599744c6-dmqzn" event={"ID":"569f0e1c-e58f-4e87-bde3-e7be4659639a","Type":"ContainerStarted","Data":"4c7e4b215beb1d463e0865db5d67c30bce7decd535274ced24856b5651616753"} Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.683809 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75599744c6-dmqzn" event={"ID":"569f0e1c-e58f-4e87-bde3-e7be4659639a","Type":"ContainerStarted","Data":"a4d123d133c5818a158baecbca329d7d1fd35c32d80f99170381689fd8534f7e"} Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.685755 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8809238d-f467-4a48-b009-15c8b974aed0","Type":"ContainerStarted","Data":"d49d97c9a969ed79389bad7a9f3619af73c623f420d9fcf018d0173695c908fc"} Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.689835 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-694b7fd487-99w9l" event={"ID":"5af85716-fd7e-45e9-8320-4db8800302a8","Type":"ContainerStarted","Data":"df625dfa320e111d2f7261b8097860d959f7d506b7125473808d8951af4edb8b"} Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.692773 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5083dbae-2860-4efd-84e5-503b3a11d99b","Type":"ContainerStarted","Data":"b513e7d8f6cdca78c4bbf0d6ee1dff4ffe512e4d9e1d9369b644275792ab4e24"} Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.733088 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.733062653 podStartE2EDuration="5.733062653s" podCreationTimestamp="2025-12-06 07:16:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:01.721142518 +0000 UTC m=+1435.176003562" watchObservedRunningTime="2025-12-06 07:17:01.733062653 +0000 UTC m=+1435.187923697" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.841360 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vlvff"] Dec 06 07:17:01 crc kubenswrapper[4945]: E1206 07:17:01.842077 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24e6d0b-d002-4d0c-8f6d-bb94de58fb90" containerName="neutron-db-sync" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.842099 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24e6d0b-d002-4d0c-8f6d-bb94de58fb90" containerName="neutron-db-sync" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.842345 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24e6d0b-d002-4d0c-8f6d-bb94de58fb90" containerName="neutron-db-sync" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.844101 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.869850 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vlvff"] Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.958720 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-svc\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.958790 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-sb\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.958843 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-config\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.958921 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgwg5\" (UniqueName: \"kubernetes.io/projected/22905fb1-0344-446f-bce3-064bed077d03-kube-api-access-zgwg5\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.959010 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-nb\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:01 crc kubenswrapper[4945]: I1206 07:17:01.959053 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-swift-storage-0\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.038379 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-577494887b-gjx4w"] Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.042415 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.044813 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-hr64m" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.045133 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.045539 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.045688 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.061783 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-nb\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.061869 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-swift-storage-0\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.061941 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-svc\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.061973 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-sb\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.062038 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-config\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.064768 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgwg5\" (UniqueName: \"kubernetes.io/projected/22905fb1-0344-446f-bce3-064bed077d03-kube-api-access-zgwg5\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.068083 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-swift-storage-0\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.068239 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-config\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.068403 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-svc\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.069028 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-nb\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.069356 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-sb\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.072145 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-577494887b-gjx4w"] Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.096047 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgwg5\" (UniqueName: \"kubernetes.io/projected/22905fb1-0344-446f-bce3-064bed077d03-kube-api-access-zgwg5\") pod \"dnsmasq-dns-77f55878d5-vlvff\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.166577 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-config\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.166797 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-combined-ca-bundle\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.166866 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m99xw\" (UniqueName: \"kubernetes.io/projected/fc779c1d-c266-43b3-95b0-8a799da93e7c-kube-api-access-m99xw\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.166974 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-httpd-config\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.167055 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-ovndb-tls-certs\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.177450 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.270487 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-config\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.270592 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-combined-ca-bundle\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.270612 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m99xw\" (UniqueName: \"kubernetes.io/projected/fc779c1d-c266-43b3-95b0-8a799da93e7c-kube-api-access-m99xw\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.270661 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-httpd-config\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.270691 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-ovndb-tls-certs\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.278085 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-ovndb-tls-certs\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.278192 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-config\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.278875 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-combined-ca-bundle\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.281214 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-httpd-config\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.300540 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m99xw\" (UniqueName: \"kubernetes.io/projected/fc779c1d-c266-43b3-95b0-8a799da93e7c-kube-api-access-m99xw\") pod \"neutron-577494887b-gjx4w\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.376484 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.734432 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75599744c6-dmqzn" event={"ID":"569f0e1c-e58f-4e87-bde3-e7be4659639a","Type":"ContainerStarted","Data":"54e516865e0c2bf291e50817a1f0a0df2b67a1322700ab4cf577d2385c08bcb2"} Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.735358 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.735399 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.745360 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8809238d-f467-4a48-b009-15c8b974aed0","Type":"ContainerStarted","Data":"b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5"} Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.747205 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-694b7fd487-99w9l" event={"ID":"5af85716-fd7e-45e9-8320-4db8800302a8","Type":"ContainerStarted","Data":"9d35634fe6ffb6779224176c397d8848048afbbeb060d73a9ed0f8871a1cd24b"} Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.789397 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-75599744c6-dmqzn" podStartSLOduration=3.789375163 podStartE2EDuration="3.789375163s" podCreationTimestamp="2025-12-06 07:16:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:02.764326432 +0000 UTC m=+1436.219187476" watchObservedRunningTime="2025-12-06 07:17:02.789375163 +0000 UTC m=+1436.244236207" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.791103 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-694b7fd487-99w9l" podStartSLOduration=3.79109485 podStartE2EDuration="3.79109485s" podCreationTimestamp="2025-12-06 07:16:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:02.785495598 +0000 UTC m=+1436.240356642" watchObservedRunningTime="2025-12-06 07:17:02.79109485 +0000 UTC m=+1436.245955894" Dec 06 07:17:02 crc kubenswrapper[4945]: I1206 07:17:02.855247 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vlvff"] Dec 06 07:17:02 crc kubenswrapper[4945]: W1206 07:17:02.857615 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22905fb1_0344_446f_bce3_064bed077d03.slice/crio-1207f71f5b15749258ee14ecd01dec8de71bd115770a413543451ce8019d3161 WatchSource:0}: Error finding container 1207f71f5b15749258ee14ecd01dec8de71bd115770a413543451ce8019d3161: Status 404 returned error can't find the container with id 1207f71f5b15749258ee14ecd01dec8de71bd115770a413543451ce8019d3161 Dec 06 07:17:03 crc kubenswrapper[4945]: I1206 07:17:03.145854 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-577494887b-gjx4w"] Dec 06 07:17:03 crc kubenswrapper[4945]: W1206 07:17:03.158781 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc779c1d_c266_43b3_95b0_8a799da93e7c.slice/crio-fe3414acc3d351c832571d8b08ef783dbd5a90259e360add7975c6b8fb5156d7 WatchSource:0}: Error finding container fe3414acc3d351c832571d8b08ef783dbd5a90259e360add7975c6b8fb5156d7: Status 404 returned error can't find the container with id fe3414acc3d351c832571d8b08ef783dbd5a90259e360add7975c6b8fb5156d7 Dec 06 07:17:03 crc kubenswrapper[4945]: I1206 07:17:03.782267 4945 generic.go:334] "Generic (PLEG): container finished" podID="22905fb1-0344-446f-bce3-064bed077d03" containerID="f75699fd19b313ef1af6fa73dc802739bc3bf9d9847e2a1244edfef2889ff773" exitCode=0 Dec 06 07:17:03 crc kubenswrapper[4945]: I1206 07:17:03.784043 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" event={"ID":"22905fb1-0344-446f-bce3-064bed077d03","Type":"ContainerDied","Data":"f75699fd19b313ef1af6fa73dc802739bc3bf9d9847e2a1244edfef2889ff773"} Dec 06 07:17:03 crc kubenswrapper[4945]: I1206 07:17:03.784107 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" event={"ID":"22905fb1-0344-446f-bce3-064bed077d03","Type":"ContainerStarted","Data":"1207f71f5b15749258ee14ecd01dec8de71bd115770a413543451ce8019d3161"} Dec 06 07:17:03 crc kubenswrapper[4945]: I1206 07:17:03.787901 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8809238d-f467-4a48-b009-15c8b974aed0","Type":"ContainerStarted","Data":"46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79"} Dec 06 07:17:03 crc kubenswrapper[4945]: I1206 07:17:03.789957 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-577494887b-gjx4w" event={"ID":"fc779c1d-c266-43b3-95b0-8a799da93e7c","Type":"ContainerStarted","Data":"fe3414acc3d351c832571d8b08ef783dbd5a90259e360add7975c6b8fb5156d7"} Dec 06 07:17:03 crc kubenswrapper[4945]: I1206 07:17:03.790713 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:03 crc kubenswrapper[4945]: I1206 07:17:03.844063 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.8440444 podStartE2EDuration="4.8440444s" podCreationTimestamp="2025-12-06 07:16:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:03.834642914 +0000 UTC m=+1437.289503968" watchObservedRunningTime="2025-12-06 07:17:03.8440444 +0000 UTC m=+1437.298905444" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.604006 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7b595c49d9-jngxp"] Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.606815 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.613199 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.613470 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.625648 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b595c49d9-jngxp"] Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.718896 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-config\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.718991 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-combined-ca-bundle\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.719097 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-public-tls-certs\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.719134 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-internal-tls-certs\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.719170 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-httpd-config\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.719212 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-ovndb-tls-certs\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.719236 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf8sb\" (UniqueName: \"kubernetes.io/projected/85ab2778-4811-4700-bca0-d55c9282699f-kube-api-access-kf8sb\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.809826 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" event={"ID":"22905fb1-0344-446f-bce3-064bed077d03","Type":"ContainerStarted","Data":"3fe5e0dcb65f6f84c1a629030e535f7279d08c3f1de45458cb189f3cd69374c1"} Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.810935 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.816809 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-577494887b-gjx4w" event={"ID":"fc779c1d-c266-43b3-95b0-8a799da93e7c","Type":"ContainerStarted","Data":"a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c"} Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.820339 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-combined-ca-bundle\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.820438 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-public-tls-certs\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.820468 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-internal-tls-certs\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.820493 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-httpd-config\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.820525 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-ovndb-tls-certs\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.820541 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf8sb\" (UniqueName: \"kubernetes.io/projected/85ab2778-4811-4700-bca0-d55c9282699f-kube-api-access-kf8sb\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.820577 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-config\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.827355 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-internal-tls-certs\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.827542 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-config\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.828268 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-ovndb-tls-certs\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.830042 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-public-tls-certs\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.834084 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-combined-ca-bundle\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.836048 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-httpd-config\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.847653 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" podStartSLOduration=3.847621147 podStartE2EDuration="3.847621147s" podCreationTimestamp="2025-12-06 07:17:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:04.835413565 +0000 UTC m=+1438.290274609" watchObservedRunningTime="2025-12-06 07:17:04.847621147 +0000 UTC m=+1438.302482191" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.858643 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf8sb\" (UniqueName: \"kubernetes.io/projected/85ab2778-4811-4700-bca0-d55c9282699f-kube-api-access-kf8sb\") pod \"neutron-7b595c49d9-jngxp\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:04 crc kubenswrapper[4945]: I1206 07:17:04.939778 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:05 crc kubenswrapper[4945]: I1206 07:17:05.831495 4945 generic.go:334] "Generic (PLEG): container finished" podID="af445462-e7c2-481c-a777-85e8566d7b71" containerID="51c9dc30540f672f0cad19373fdd58c9f35dc611b2c4f98ab2d185612f7a7af6" exitCode=0 Dec 06 07:17:05 crc kubenswrapper[4945]: I1206 07:17:05.831540 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cks9w" event={"ID":"af445462-e7c2-481c-a777-85e8566d7b71","Type":"ContainerDied","Data":"51c9dc30540f672f0cad19373fdd58c9f35dc611b2c4f98ab2d185612f7a7af6"} Dec 06 07:17:07 crc kubenswrapper[4945]: I1206 07:17:07.206664 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 07:17:07 crc kubenswrapper[4945]: I1206 07:17:07.207070 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 07:17:07 crc kubenswrapper[4945]: I1206 07:17:07.249631 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 07:17:07 crc kubenswrapper[4945]: I1206 07:17:07.254773 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 07:17:07 crc kubenswrapper[4945]: I1206 07:17:07.859225 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 07:17:07 crc kubenswrapper[4945]: I1206 07:17:07.859360 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 07:17:08 crc kubenswrapper[4945]: I1206 07:17:08.796050 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:17:08 crc kubenswrapper[4945]: I1206 07:17:08.796477 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:17:09 crc kubenswrapper[4945]: I1206 07:17:09.998022 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 07:17:09 crc kubenswrapper[4945]: I1206 07:17:09.998453 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.377389 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.377436 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.385903 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cks9w" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.423099 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.434273 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.457925 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.472213 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-combined-ca-bundle\") pod \"af445462-e7c2-481c-a777-85e8566d7b71\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.472313 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-db-sync-config-data\") pod \"af445462-e7c2-481c-a777-85e8566d7b71\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.472357 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sflxq\" (UniqueName: \"kubernetes.io/projected/af445462-e7c2-481c-a777-85e8566d7b71-kube-api-access-sflxq\") pod \"af445462-e7c2-481c-a777-85e8566d7b71\" (UID: \"af445462-e7c2-481c-a777-85e8566d7b71\") " Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.480255 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af445462-e7c2-481c-a777-85e8566d7b71-kube-api-access-sflxq" (OuterVolumeSpecName: "kube-api-access-sflxq") pod "af445462-e7c2-481c-a777-85e8566d7b71" (UID: "af445462-e7c2-481c-a777-85e8566d7b71"). InnerVolumeSpecName "kube-api-access-sflxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.496050 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "af445462-e7c2-481c-a777-85e8566d7b71" (UID: "af445462-e7c2-481c-a777-85e8566d7b71"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.524457 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af445462-e7c2-481c-a777-85e8566d7b71" (UID: "af445462-e7c2-481c-a777-85e8566d7b71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.574364 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.574401 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af445462-e7c2-481c-a777-85e8566d7b71-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.574412 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sflxq\" (UniqueName: \"kubernetes.io/projected/af445462-e7c2-481c-a777-85e8566d7b71-kube-api-access-sflxq\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.888535 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cks9w" event={"ID":"af445462-e7c2-481c-a777-85e8566d7b71","Type":"ContainerDied","Data":"0042159a2556bbacefdcefda2eda8a4872bb974b7dc33ad94ea7f6101570355b"} Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.888657 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0042159a2556bbacefdcefda2eda8a4872bb974b7dc33ad94ea7f6101570355b" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.889037 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cks9w" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.889082 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:10 crc kubenswrapper[4945]: I1206 07:17:10.889101 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.676608 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-687fd77bc7-9rms8"] Dec 06 07:17:11 crc kubenswrapper[4945]: E1206 07:17:11.677616 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af445462-e7c2-481c-a777-85e8566d7b71" containerName="barbican-db-sync" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.677634 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af445462-e7c2-481c-a777-85e8566d7b71" containerName="barbican-db-sync" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.677860 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af445462-e7c2-481c-a777-85e8566d7b71" containerName="barbican-db-sync" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.678872 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.682811 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.683049 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kgdxv" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.686633 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.704541 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-logs\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.704685 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data-custom\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.704737 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxbsj\" (UniqueName: \"kubernetes.io/projected/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-kube-api-access-cxbsj\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.704815 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-combined-ca-bundle\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.704867 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.725353 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7db96d494-qst22"] Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.727252 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.731634 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.756696 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-687fd77bc7-9rms8"] Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.766916 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7db96d494-qst22"] Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.808970 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.809069 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data-custom\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.809206 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/966c9b21-65c4-4d2a-af0f-0041a9bfea08-logs\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.809239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxbsj\" (UniqueName: \"kubernetes.io/projected/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-kube-api-access-cxbsj\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.820163 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-combined-ca-bundle\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.820327 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.820393 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-combined-ca-bundle\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.820572 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n22t\" (UniqueName: \"kubernetes.io/projected/966c9b21-65c4-4d2a-af0f-0041a9bfea08-kube-api-access-2n22t\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.820625 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-logs\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.820663 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data-custom\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.821545 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-logs\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.828438 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-combined-ca-bundle\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.844033 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data-custom\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.852689 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.853005 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vlvff"] Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.853259 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" podUID="22905fb1-0344-446f-bce3-064bed077d03" containerName="dnsmasq-dns" containerID="cri-o://3fe5e0dcb65f6f84c1a629030e535f7279d08c3f1de45458cb189f3cd69374c1" gracePeriod=10 Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.854875 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxbsj\" (UniqueName: \"kubernetes.io/projected/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-kube-api-access-cxbsj\") pod \"barbican-worker-687fd77bc7-9rms8\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.858402 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.926328 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.926412 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/966c9b21-65c4-4d2a-af0f-0041a9bfea08-logs\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.926504 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-combined-ca-bundle\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.926570 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n22t\" (UniqueName: \"kubernetes.io/projected/966c9b21-65c4-4d2a-af0f-0041a9bfea08-kube-api-access-2n22t\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.926615 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data-custom\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.927575 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/966c9b21-65c4-4d2a-af0f-0041a9bfea08-logs\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:11 crc kubenswrapper[4945]: I1206 07:17:11.980550 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n22t\" (UniqueName: \"kubernetes.io/projected/966c9b21-65c4-4d2a-af0f-0041a9bfea08-kube-api-access-2n22t\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.008853 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-7kgt2"] Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.030738 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data-custom\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.030778 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.035705 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-combined-ca-bundle\") pod \"barbican-keystone-listener-7db96d494-qst22\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.053743 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.073593 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-7kgt2"] Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.078332 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.103092 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.111832 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b595c49d9-jngxp"] Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.123381 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5899f79f4b-lsl4d"] Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.125209 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.128013 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.143374 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5899f79f4b-lsl4d"] Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.148713 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-svc\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.148898 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-swift-storage-0\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.148945 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmvjn\" (UniqueName: \"kubernetes.io/projected/b75c5b75-ef89-4898-aca5-7c3bd94cda90-kube-api-access-xmvjn\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.148967 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-sb\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.149056 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-config\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.149106 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-nb\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.179289 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" podUID="22905fb1-0344-446f-bce3-064bed077d03" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: connect: connection refused" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250503 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p29lc\" (UniqueName: \"kubernetes.io/projected/9d840c09-0868-449e-9699-9b84aa8e0fce-kube-api-access-p29lc\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250583 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data-custom\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250645 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-combined-ca-bundle\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250678 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250747 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-swift-storage-0\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250797 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmvjn\" (UniqueName: \"kubernetes.io/projected/b75c5b75-ef89-4898-aca5-7c3bd94cda90-kube-api-access-xmvjn\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250828 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-sb\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250862 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d840c09-0868-449e-9699-9b84aa8e0fce-logs\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250894 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-config\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250921 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-nb\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.250961 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-svc\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.251961 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-svc\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.252449 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-config\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.252456 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-sb\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.254817 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-swift-storage-0\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.254900 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-nb\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.270665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmvjn\" (UniqueName: \"kubernetes.io/projected/b75c5b75-ef89-4898-aca5-7c3bd94cda90-kube-api-access-xmvjn\") pod \"dnsmasq-dns-687dbb56f-7kgt2\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.352357 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d840c09-0868-449e-9699-9b84aa8e0fce-logs\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.352484 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p29lc\" (UniqueName: \"kubernetes.io/projected/9d840c09-0868-449e-9699-9b84aa8e0fce-kube-api-access-p29lc\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.352517 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data-custom\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.352540 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-combined-ca-bundle\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.352564 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.353436 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d840c09-0868-449e-9699-9b84aa8e0fce-logs\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.358758 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-combined-ca-bundle\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.359382 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.359911 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data-custom\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.374699 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p29lc\" (UniqueName: \"kubernetes.io/projected/9d840c09-0868-449e-9699-9b84aa8e0fce-kube-api-access-p29lc\") pod \"barbican-api-5899f79f4b-lsl4d\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.430040 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.502798 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.980899 4945 generic.go:334] "Generic (PLEG): container finished" podID="22905fb1-0344-446f-bce3-064bed077d03" containerID="3fe5e0dcb65f6f84c1a629030e535f7279d08c3f1de45458cb189f3cd69374c1" exitCode=0 Dec 06 07:17:12 crc kubenswrapper[4945]: I1206 07:17:12.980946 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" event={"ID":"22905fb1-0344-446f-bce3-064bed077d03","Type":"ContainerDied","Data":"3fe5e0dcb65f6f84c1a629030e535f7279d08c3f1de45458cb189f3cd69374c1"} Dec 06 07:17:13 crc kubenswrapper[4945]: W1206 07:17:13.182412 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85ab2778_4811_4700_bca0_d55c9282699f.slice/crio-29d2871cbbd51f2c03441cdf4a01555e9f6c5623844918cfeecdd50cd701d0b8 WatchSource:0}: Error finding container 29d2871cbbd51f2c03441cdf4a01555e9f6c5623844918cfeecdd50cd701d0b8: Status 404 returned error can't find the container with id 29d2871cbbd51f2c03441cdf4a01555e9f6c5623844918cfeecdd50cd701d0b8 Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.700704 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.728050 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.728215 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.795696 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-sb\") pod \"22905fb1-0344-446f-bce3-064bed077d03\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.795768 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-swift-storage-0\") pod \"22905fb1-0344-446f-bce3-064bed077d03\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.795806 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-svc\") pod \"22905fb1-0344-446f-bce3-064bed077d03\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.795890 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-nb\") pod \"22905fb1-0344-446f-bce3-064bed077d03\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.795965 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgwg5\" (UniqueName: \"kubernetes.io/projected/22905fb1-0344-446f-bce3-064bed077d03-kube-api-access-zgwg5\") pod \"22905fb1-0344-446f-bce3-064bed077d03\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.795985 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-config\") pod \"22905fb1-0344-446f-bce3-064bed077d03\" (UID: \"22905fb1-0344-446f-bce3-064bed077d03\") " Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.815328 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22905fb1-0344-446f-bce3-064bed077d03-kube-api-access-zgwg5" (OuterVolumeSpecName: "kube-api-access-zgwg5") pod "22905fb1-0344-446f-bce3-064bed077d03" (UID: "22905fb1-0344-446f-bce3-064bed077d03"). InnerVolumeSpecName "kube-api-access-zgwg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.898388 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgwg5\" (UniqueName: \"kubernetes.io/projected/22905fb1-0344-446f-bce3-064bed077d03-kube-api-access-zgwg5\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:13 crc kubenswrapper[4945]: I1206 07:17:13.905498 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-7kgt2"] Dec 06 07:17:13 crc kubenswrapper[4945]: W1206 07:17:13.966293 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb75c5b75_ef89_4898_aca5_7c3bd94cda90.slice/crio-25e52c74a72794d95143f897ac67d7b895460dcd93696d15d4b17da9b6ee85f2 WatchSource:0}: Error finding container 25e52c74a72794d95143f897ac67d7b895460dcd93696d15d4b17da9b6ee85f2: Status 404 returned error can't find the container with id 25e52c74a72794d95143f897ac67d7b895460dcd93696d15d4b17da9b6ee85f2 Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.018702 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-577494887b-gjx4w" event={"ID":"fc779c1d-c266-43b3-95b0-8a799da93e7c","Type":"ContainerStarted","Data":"bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48"} Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.018872 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.024525 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" event={"ID":"b75c5b75-ef89-4898-aca5-7c3bd94cda90","Type":"ContainerStarted","Data":"25e52c74a72794d95143f897ac67d7b895460dcd93696d15d4b17da9b6ee85f2"} Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.029345 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" event={"ID":"22905fb1-0344-446f-bce3-064bed077d03","Type":"ContainerDied","Data":"1207f71f5b15749258ee14ecd01dec8de71bd115770a413543451ce8019d3161"} Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.029396 4945 scope.go:117] "RemoveContainer" containerID="3fe5e0dcb65f6f84c1a629030e535f7279d08c3f1de45458cb189f3cd69374c1" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.029443 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-vlvff" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.059956 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b595c49d9-jngxp" event={"ID":"85ab2778-4811-4700-bca0-d55c9282699f","Type":"ContainerStarted","Data":"29d2871cbbd51f2c03441cdf4a01555e9f6c5623844918cfeecdd50cd701d0b8"} Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.078172 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.082594 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-577494887b-gjx4w" podStartSLOduration=12.082571112 podStartE2EDuration="12.082571112s" podCreationTimestamp="2025-12-06 07:17:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:14.059172945 +0000 UTC m=+1447.514033999" watchObservedRunningTime="2025-12-06 07:17:14.082571112 +0000 UTC m=+1447.537432146" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.087427 4945 scope.go:117] "RemoveContainer" containerID="f75699fd19b313ef1af6fa73dc802739bc3bf9d9847e2a1244edfef2889ff773" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.263599 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5899f79f4b-lsl4d"] Dec 06 07:17:14 crc kubenswrapper[4945]: E1206 07:17:14.267691 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.316442 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7db96d494-qst22"] Dec 06 07:17:14 crc kubenswrapper[4945]: W1206 07:17:14.331470 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d840c09_0868_449e_9699_9b84aa8e0fce.slice/crio-800d7c1faa962d8af183da7c56ddcb721fb6ffd71ca900156565b59649d7d43d WatchSource:0}: Error finding container 800d7c1faa962d8af183da7c56ddcb721fb6ffd71ca900156565b59649d7d43d: Status 404 returned error can't find the container with id 800d7c1faa962d8af183da7c56ddcb721fb6ffd71ca900156565b59649d7d43d Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.333162 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-687fd77bc7-9rms8"] Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.342528 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "22905fb1-0344-446f-bce3-064bed077d03" (UID: "22905fb1-0344-446f-bce3-064bed077d03"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.388372 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "22905fb1-0344-446f-bce3-064bed077d03" (UID: "22905fb1-0344-446f-bce3-064bed077d03"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.388964 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "22905fb1-0344-446f-bce3-064bed077d03" (UID: "22905fb1-0344-446f-bce3-064bed077d03"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.394948 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "22905fb1-0344-446f-bce3-064bed077d03" (UID: "22905fb1-0344-446f-bce3-064bed077d03"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.431630 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.431670 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.431684 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.431694 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.434566 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-config" (OuterVolumeSpecName: "config") pod "22905fb1-0344-446f-bce3-064bed077d03" (UID: "22905fb1-0344-446f-bce3-064bed077d03"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.533822 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22905fb1-0344-446f-bce3-064bed077d03-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.716611 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vlvff"] Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.725339 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vlvff"] Dec 06 07:17:14 crc kubenswrapper[4945]: I1206 07:17:14.972536 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22905fb1-0344-446f-bce3-064bed077d03" path="/var/lib/kubelet/pods/22905fb1-0344-446f-bce3-064bed077d03/volumes" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.083040 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hr6hb" event={"ID":"aa503636-fff0-4195-9364-91c2f03e5c9b","Type":"ContainerStarted","Data":"3df7ad5bd72c70655a8062719a5e2173ccd640642cd65ada9584c007120552b2"} Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.087700 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5899f79f4b-lsl4d" event={"ID":"9d840c09-0868-449e-9699-9b84aa8e0fce","Type":"ContainerStarted","Data":"023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822"} Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.087864 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5899f79f4b-lsl4d" event={"ID":"9d840c09-0868-449e-9699-9b84aa8e0fce","Type":"ContainerStarted","Data":"800d7c1faa962d8af183da7c56ddcb721fb6ffd71ca900156565b59649d7d43d"} Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.099113 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-687fd77bc7-9rms8" event={"ID":"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac","Type":"ContainerStarted","Data":"3aaf92a05d038b9022c400636319ff4cd912351991d354de0369dc8c14fb8955"} Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.104067 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7db96d494-qst22" event={"ID":"966c9b21-65c4-4d2a-af0f-0041a9bfea08","Type":"ContainerStarted","Data":"e5ec440e1b2982d8d5e5efdec9ff243026d7ad0b6f7413eee72b25f1c1bcbb7c"} Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.105856 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b595c49d9-jngxp" event={"ID":"85ab2778-4811-4700-bca0-d55c9282699f","Type":"ContainerStarted","Data":"9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6"} Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.105883 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b595c49d9-jngxp" event={"ID":"85ab2778-4811-4700-bca0-d55c9282699f","Type":"ContainerStarted","Data":"915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea"} Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.107153 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.111740 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-hr6hb" podStartSLOduration=4.043425257 podStartE2EDuration="53.111715424s" podCreationTimestamp="2025-12-06 07:16:22 +0000 UTC" firstStartedPulling="2025-12-06 07:16:24.115454018 +0000 UTC m=+1397.570315062" lastFinishedPulling="2025-12-06 07:17:13.183744185 +0000 UTC m=+1446.638605229" observedRunningTime="2025-12-06 07:17:15.102941286 +0000 UTC m=+1448.557802350" watchObservedRunningTime="2025-12-06 07:17:15.111715424 +0000 UTC m=+1448.566576468" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.120761 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e525fb3-9948-4038-afce-b863a4ca5206","Type":"ContainerStarted","Data":"de48e720232535cebe13a20fa073ca681801633b91ea03746596672748416839"} Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.121097 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="ceilometer-notification-agent" containerID="cri-o://fdee576b16798f10cbf0d6df65b22dc489c5de6fce0d46d09574de5ceaf78ca7" gracePeriod=30 Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.121151 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="proxy-httpd" containerID="cri-o://de48e720232535cebe13a20fa073ca681801633b91ea03746596672748416839" gracePeriod=30 Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.121182 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="sg-core" containerID="cri-o://2ddc90f531ce8123229116bd612f56498d8293ed1cb1cd9acab8b626c6700c6c" gracePeriod=30 Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.121205 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.151429 4945 generic.go:334] "Generic (PLEG): container finished" podID="b75c5b75-ef89-4898-aca5-7c3bd94cda90" containerID="14844dfa9e120ac477947241791eb4b53a707c2e73c9bd37e9218941aae58d93" exitCode=0 Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.152367 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" event={"ID":"b75c5b75-ef89-4898-aca5-7c3bd94cda90","Type":"ContainerDied","Data":"14844dfa9e120ac477947241791eb4b53a707c2e73c9bd37e9218941aae58d93"} Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.162474 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7b595c49d9-jngxp" podStartSLOduration=11.162435974 podStartE2EDuration="11.162435974s" podCreationTimestamp="2025-12-06 07:17:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:15.135233164 +0000 UTC m=+1448.590094208" watchObservedRunningTime="2025-12-06 07:17:15.162435974 +0000 UTC m=+1448.617297018" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.919561 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f9668d888-8gtj6"] Dec 06 07:17:15 crc kubenswrapper[4945]: E1206 07:17:15.920230 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22905fb1-0344-446f-bce3-064bed077d03" containerName="dnsmasq-dns" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.920248 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="22905fb1-0344-446f-bce3-064bed077d03" containerName="dnsmasq-dns" Dec 06 07:17:15 crc kubenswrapper[4945]: E1206 07:17:15.920271 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22905fb1-0344-446f-bce3-064bed077d03" containerName="init" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.920407 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="22905fb1-0344-446f-bce3-064bed077d03" containerName="init" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.924891 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="22905fb1-0344-446f-bce3-064bed077d03" containerName="dnsmasq-dns" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.926172 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.929175 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.929496 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 06 07:17:15 crc kubenswrapper[4945]: I1206 07:17:15.939251 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f9668d888-8gtj6"] Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.002232 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-public-tls-certs\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.002559 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-internal-tls-certs\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.002621 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-combined-ca-bundle\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.002748 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8hw2\" (UniqueName: \"kubernetes.io/projected/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-kube-api-access-x8hw2\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.002991 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.003036 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data-custom\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.003108 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-logs\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.104628 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-public-tls-certs\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.104707 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-internal-tls-certs\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.104743 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-combined-ca-bundle\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.104803 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8hw2\" (UniqueName: \"kubernetes.io/projected/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-kube-api-access-x8hw2\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.104894 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.104920 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data-custom\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.104935 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-logs\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.105421 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-logs\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.113020 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-public-tls-certs\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.114142 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-internal-tls-certs\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.115575 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.119901 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data-custom\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.120097 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-combined-ca-bundle\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.130595 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8hw2\" (UniqueName: \"kubernetes.io/projected/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-kube-api-access-x8hw2\") pod \"barbican-api-5f9668d888-8gtj6\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.181354 4945 generic.go:334] "Generic (PLEG): container finished" podID="4e525fb3-9948-4038-afce-b863a4ca5206" containerID="de48e720232535cebe13a20fa073ca681801633b91ea03746596672748416839" exitCode=0 Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.181395 4945 generic.go:334] "Generic (PLEG): container finished" podID="4e525fb3-9948-4038-afce-b863a4ca5206" containerID="2ddc90f531ce8123229116bd612f56498d8293ed1cb1cd9acab8b626c6700c6c" exitCode=2 Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.181443 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e525fb3-9948-4038-afce-b863a4ca5206","Type":"ContainerDied","Data":"de48e720232535cebe13a20fa073ca681801633b91ea03746596672748416839"} Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.181477 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e525fb3-9948-4038-afce-b863a4ca5206","Type":"ContainerDied","Data":"2ddc90f531ce8123229116bd612f56498d8293ed1cb1cd9acab8b626c6700c6c"} Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.188586 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" event={"ID":"b75c5b75-ef89-4898-aca5-7c3bd94cda90","Type":"ContainerStarted","Data":"1f2f55eba30b22190ba637ec820a6dc48392790ff0937bf474cf5b424bdb9e78"} Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.190025 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.214509 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5899f79f4b-lsl4d" event={"ID":"9d840c09-0868-449e-9699-9b84aa8e0fce","Type":"ContainerStarted","Data":"189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a"} Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.215319 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.234461 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" podStartSLOduration=5.234437703 podStartE2EDuration="5.234437703s" podCreationTimestamp="2025-12-06 07:17:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:16.209910716 +0000 UTC m=+1449.664771780" watchObservedRunningTime="2025-12-06 07:17:16.234437703 +0000 UTC m=+1449.689298747" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.246740 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:16 crc kubenswrapper[4945]: I1206 07:17:16.248116 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5899f79f4b-lsl4d" podStartSLOduration=4.248094064 podStartE2EDuration="4.248094064s" podCreationTimestamp="2025-12-06 07:17:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:16.239824189 +0000 UTC m=+1449.694685243" watchObservedRunningTime="2025-12-06 07:17:16.248094064 +0000 UTC m=+1449.702955108" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.244320 4945 generic.go:334] "Generic (PLEG): container finished" podID="4e525fb3-9948-4038-afce-b863a4ca5206" containerID="fdee576b16798f10cbf0d6df65b22dc489c5de6fce0d46d09574de5ceaf78ca7" exitCode=0 Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.244417 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e525fb3-9948-4038-afce-b863a4ca5206","Type":"ContainerDied","Data":"fdee576b16798f10cbf0d6df65b22dc489c5de6fce0d46d09574de5ceaf78ca7"} Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.244965 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e525fb3-9948-4038-afce-b863a4ca5206","Type":"ContainerDied","Data":"9475d101b613b59f4757a824d9dc3749b65b5b2ae7a9fa19e7510f7c41bc2f45"} Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.244983 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9475d101b613b59f4757a824d9dc3749b65b5b2ae7a9fa19e7510f7c41bc2f45" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.246164 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.350066 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.535355 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-sg-core-conf-yaml\") pod \"4e525fb3-9948-4038-afce-b863a4ca5206\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.535446 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-config-data\") pod \"4e525fb3-9948-4038-afce-b863a4ca5206\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.535532 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-combined-ca-bundle\") pod \"4e525fb3-9948-4038-afce-b863a4ca5206\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.535624 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-run-httpd\") pod \"4e525fb3-9948-4038-afce-b863a4ca5206\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.535735 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n8mq\" (UniqueName: \"kubernetes.io/projected/4e525fb3-9948-4038-afce-b863a4ca5206-kube-api-access-6n8mq\") pod \"4e525fb3-9948-4038-afce-b863a4ca5206\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.535831 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-scripts\") pod \"4e525fb3-9948-4038-afce-b863a4ca5206\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.535867 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-log-httpd\") pod \"4e525fb3-9948-4038-afce-b863a4ca5206\" (UID: \"4e525fb3-9948-4038-afce-b863a4ca5206\") " Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.536913 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4e525fb3-9948-4038-afce-b863a4ca5206" (UID: "4e525fb3-9948-4038-afce-b863a4ca5206"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.537435 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4e525fb3-9948-4038-afce-b863a4ca5206" (UID: "4e525fb3-9948-4038-afce-b863a4ca5206"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.543204 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e525fb3-9948-4038-afce-b863a4ca5206-kube-api-access-6n8mq" (OuterVolumeSpecName: "kube-api-access-6n8mq") pod "4e525fb3-9948-4038-afce-b863a4ca5206" (UID: "4e525fb3-9948-4038-afce-b863a4ca5206"). InnerVolumeSpecName "kube-api-access-6n8mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.550849 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-scripts" (OuterVolumeSpecName: "scripts") pod "4e525fb3-9948-4038-afce-b863a4ca5206" (UID: "4e525fb3-9948-4038-afce-b863a4ca5206"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.577244 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4e525fb3-9948-4038-afce-b863a4ca5206" (UID: "4e525fb3-9948-4038-afce-b863a4ca5206"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.617488 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e525fb3-9948-4038-afce-b863a4ca5206" (UID: "4e525fb3-9948-4038-afce-b863a4ca5206"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.638757 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n8mq\" (UniqueName: \"kubernetes.io/projected/4e525fb3-9948-4038-afce-b863a4ca5206-kube-api-access-6n8mq\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.638795 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.638806 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.638819 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.638833 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.638845 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e525fb3-9948-4038-afce-b863a4ca5206-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.654574 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-config-data" (OuterVolumeSpecName: "config-data") pod "4e525fb3-9948-4038-afce-b863a4ca5206" (UID: "4e525fb3-9948-4038-afce-b863a4ca5206"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.657863 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f9668d888-8gtj6"] Dec 06 07:17:17 crc kubenswrapper[4945]: I1206 07:17:17.740845 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e525fb3-9948-4038-afce-b863a4ca5206-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.266863 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9668d888-8gtj6" event={"ID":"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975","Type":"ContainerStarted","Data":"afba2827fdc58ee118d91cef305ad6d598b1747d9ddf4810bfde53456fdf7417"} Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.266946 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9668d888-8gtj6" event={"ID":"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975","Type":"ContainerStarted","Data":"1c796787c79166cd1f6bdca3eb60a502cd2b2fc6bd58628b9bff08d4c77ca7c4"} Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.266959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9668d888-8gtj6" event={"ID":"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975","Type":"ContainerStarted","Data":"67db1ce53a734cdc8d1cb41de803b70c6fd26d779a4527d1909b5ec2a833c239"} Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.267246 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.273349 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-687fd77bc7-9rms8" event={"ID":"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac","Type":"ContainerStarted","Data":"8f06adda9c4f122a20f297ca0d7ccd2c0349a987c76bc355daf2b4ea17945592"} Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.273468 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-687fd77bc7-9rms8" event={"ID":"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac","Type":"ContainerStarted","Data":"2326a98863fbfe1d3fcf0bf3e1db275908b04d5165d45b924f0b985263d8c616"} Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.292228 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f9668d888-8gtj6" podStartSLOduration=3.292207873 podStartE2EDuration="3.292207873s" podCreationTimestamp="2025-12-06 07:17:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:18.288228164 +0000 UTC m=+1451.743089208" watchObservedRunningTime="2025-12-06 07:17:18.292207873 +0000 UTC m=+1451.747068917" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.294568 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7db96d494-qst22" event={"ID":"966c9b21-65c4-4d2a-af0f-0041a9bfea08","Type":"ContainerStarted","Data":"c997070637d077fb0bd2b5c6e603ae66c9789fb87271122eac47abaf2b27318e"} Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.294630 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7db96d494-qst22" event={"ID":"966c9b21-65c4-4d2a-af0f-0041a9bfea08","Type":"ContainerStarted","Data":"61c8d1dfdac1f1aac44c2ab0b8448e2373849c62730037840a5d71971f8623dc"} Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.295133 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.329361 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-687fd77bc7-9rms8" podStartSLOduration=4.518094551 podStartE2EDuration="7.329343963s" podCreationTimestamp="2025-12-06 07:17:11 +0000 UTC" firstStartedPulling="2025-12-06 07:17:14.368780169 +0000 UTC m=+1447.823641213" lastFinishedPulling="2025-12-06 07:17:17.180029581 +0000 UTC m=+1450.634890625" observedRunningTime="2025-12-06 07:17:18.328914391 +0000 UTC m=+1451.783775445" watchObservedRunningTime="2025-12-06 07:17:18.329343963 +0000 UTC m=+1451.784205007" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.375782 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7db96d494-qst22" podStartSLOduration=4.553129155 podStartE2EDuration="7.375760226s" podCreationTimestamp="2025-12-06 07:17:11 +0000 UTC" firstStartedPulling="2025-12-06 07:17:14.354124261 +0000 UTC m=+1447.808985315" lastFinishedPulling="2025-12-06 07:17:17.176755342 +0000 UTC m=+1450.631616386" observedRunningTime="2025-12-06 07:17:18.354808316 +0000 UTC m=+1451.809669370" watchObservedRunningTime="2025-12-06 07:17:18.375760226 +0000 UTC m=+1451.830621260" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.466373 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.479180 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.511207 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:18 crc kubenswrapper[4945]: E1206 07:17:18.512458 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="ceilometer-notification-agent" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.512485 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="ceilometer-notification-agent" Dec 06 07:17:18 crc kubenswrapper[4945]: E1206 07:17:18.512506 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="proxy-httpd" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.512513 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="proxy-httpd" Dec 06 07:17:18 crc kubenswrapper[4945]: E1206 07:17:18.512555 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="sg-core" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.512562 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="sg-core" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.512970 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="proxy-httpd" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.512993 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="ceilometer-notification-agent" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.513018 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" containerName="sg-core" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.539526 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.539689 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.544883 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.557323 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.595443 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-log-httpd\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.595501 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.595530 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-run-httpd\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.595599 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-scripts\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.595617 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-config-data\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.595641 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.595690 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmqcw\" (UniqueName: \"kubernetes.io/projected/72e9a7b2-b482-4130-819a-95ac1106e218-kube-api-access-pmqcw\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.696997 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmqcw\" (UniqueName: \"kubernetes.io/projected/72e9a7b2-b482-4130-819a-95ac1106e218-kube-api-access-pmqcw\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.697616 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-log-httpd\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.697716 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.697807 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-run-httpd\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.697925 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-scripts\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.697995 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-config-data\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.698076 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.699117 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-log-httpd\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.699153 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-run-httpd\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.705362 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.712383 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-config-data\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.713538 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-scripts\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.713974 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.718365 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmqcw\" (UniqueName: \"kubernetes.io/projected/72e9a7b2-b482-4130-819a-95ac1106e218-kube-api-access-pmqcw\") pod \"ceilometer-0\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.867305 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:17:18 crc kubenswrapper[4945]: I1206 07:17:18.979390 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e525fb3-9948-4038-afce-b863a4ca5206" path="/var/lib/kubelet/pods/4e525fb3-9948-4038-afce-b863a4ca5206/volumes" Dec 06 07:17:19 crc kubenswrapper[4945]: I1206 07:17:19.302016 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:19 crc kubenswrapper[4945]: I1206 07:17:19.428777 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:20 crc kubenswrapper[4945]: I1206 07:17:20.313308 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72e9a7b2-b482-4130-819a-95ac1106e218","Type":"ContainerStarted","Data":"2139e668db9a92f5f7706eff1e6c504a57c97c94612cadb38beac36f2fa9ed10"} Dec 06 07:17:20 crc kubenswrapper[4945]: I1206 07:17:20.316663 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72e9a7b2-b482-4130-819a-95ac1106e218","Type":"ContainerStarted","Data":"b067a2c52527c1fa782f2408d497d11103b1b4f6a63115a9e188fbf38a9ed35a"} Dec 06 07:17:21 crc kubenswrapper[4945]: I1206 07:17:21.324609 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72e9a7b2-b482-4130-819a-95ac1106e218","Type":"ContainerStarted","Data":"3ab75bae5874506e9ae523bd85a3d001e584020c211168290d995e46abd09b61"} Dec 06 07:17:21 crc kubenswrapper[4945]: I1206 07:17:21.326856 4945 generic.go:334] "Generic (PLEG): container finished" podID="aa503636-fff0-4195-9364-91c2f03e5c9b" containerID="3df7ad5bd72c70655a8062719a5e2173ccd640642cd65ada9584c007120552b2" exitCode=0 Dec 06 07:17:21 crc kubenswrapper[4945]: I1206 07:17:21.326881 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hr6hb" event={"ID":"aa503636-fff0-4195-9364-91c2f03e5c9b","Type":"ContainerDied","Data":"3df7ad5bd72c70655a8062719a5e2173ccd640642cd65ada9584c007120552b2"} Dec 06 07:17:22 crc kubenswrapper[4945]: I1206 07:17:22.346625 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72e9a7b2-b482-4130-819a-95ac1106e218","Type":"ContainerStarted","Data":"df489374460e08de26f48143dd596d770a90fe2c61e81c40c9410a4bcd4d5df6"} Dec 06 07:17:22 crc kubenswrapper[4945]: I1206 07:17:22.433458 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:22 crc kubenswrapper[4945]: I1206 07:17:22.592824 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-khdrt"] Dec 06 07:17:22 crc kubenswrapper[4945]: I1206 07:17:22.593286 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" podUID="6cdd3739-5157-4344-acfc-8ff075bc91b8" containerName="dnsmasq-dns" containerID="cri-o://f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a" gracePeriod=10 Dec 06 07:17:22 crc kubenswrapper[4945]: I1206 07:17:22.917791 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.042968 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-combined-ca-bundle\") pod \"aa503636-fff0-4195-9364-91c2f03e5c9b\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.043342 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aa503636-fff0-4195-9364-91c2f03e5c9b-etc-machine-id\") pod \"aa503636-fff0-4195-9364-91c2f03e5c9b\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.043379 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-scripts\") pod \"aa503636-fff0-4195-9364-91c2f03e5c9b\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.043402 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hphb\" (UniqueName: \"kubernetes.io/projected/aa503636-fff0-4195-9364-91c2f03e5c9b-kube-api-access-2hphb\") pod \"aa503636-fff0-4195-9364-91c2f03e5c9b\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.043504 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-db-sync-config-data\") pod \"aa503636-fff0-4195-9364-91c2f03e5c9b\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.043569 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-config-data\") pod \"aa503636-fff0-4195-9364-91c2f03e5c9b\" (UID: \"aa503636-fff0-4195-9364-91c2f03e5c9b\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.053932 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa503636-fff0-4195-9364-91c2f03e5c9b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "aa503636-fff0-4195-9364-91c2f03e5c9b" (UID: "aa503636-fff0-4195-9364-91c2f03e5c9b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.075680 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa503636-fff0-4195-9364-91c2f03e5c9b-kube-api-access-2hphb" (OuterVolumeSpecName: "kube-api-access-2hphb") pod "aa503636-fff0-4195-9364-91c2f03e5c9b" (UID: "aa503636-fff0-4195-9364-91c2f03e5c9b"). InnerVolumeSpecName "kube-api-access-2hphb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.075810 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-scripts" (OuterVolumeSpecName: "scripts") pod "aa503636-fff0-4195-9364-91c2f03e5c9b" (UID: "aa503636-fff0-4195-9364-91c2f03e5c9b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.104507 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "aa503636-fff0-4195-9364-91c2f03e5c9b" (UID: "aa503636-fff0-4195-9364-91c2f03e5c9b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.120955 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa503636-fff0-4195-9364-91c2f03e5c9b" (UID: "aa503636-fff0-4195-9364-91c2f03e5c9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.145640 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.145668 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.145678 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aa503636-fff0-4195-9364-91c2f03e5c9b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.145687 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.145698 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hphb\" (UniqueName: \"kubernetes.io/projected/aa503636-fff0-4195-9364-91c2f03e5c9b-kube-api-access-2hphb\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.167908 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.181415 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-config-data" (OuterVolumeSpecName: "config-data") pod "aa503636-fff0-4195-9364-91c2f03e5c9b" (UID: "aa503636-fff0-4195-9364-91c2f03e5c9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.247225 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa503636-fff0-4195-9364-91c2f03e5c9b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.348389 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-nb\") pod \"6cdd3739-5157-4344-acfc-8ff075bc91b8\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.348632 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-swift-storage-0\") pod \"6cdd3739-5157-4344-acfc-8ff075bc91b8\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.348661 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-svc\") pod \"6cdd3739-5157-4344-acfc-8ff075bc91b8\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.348701 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-config\") pod \"6cdd3739-5157-4344-acfc-8ff075bc91b8\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.348751 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8rj5\" (UniqueName: \"kubernetes.io/projected/6cdd3739-5157-4344-acfc-8ff075bc91b8-kube-api-access-z8rj5\") pod \"6cdd3739-5157-4344-acfc-8ff075bc91b8\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.348806 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-sb\") pod \"6cdd3739-5157-4344-acfc-8ff075bc91b8\" (UID: \"6cdd3739-5157-4344-acfc-8ff075bc91b8\") " Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.353867 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cdd3739-5157-4344-acfc-8ff075bc91b8-kube-api-access-z8rj5" (OuterVolumeSpecName: "kube-api-access-z8rj5") pod "6cdd3739-5157-4344-acfc-8ff075bc91b8" (UID: "6cdd3739-5157-4344-acfc-8ff075bc91b8"). InnerVolumeSpecName "kube-api-access-z8rj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.365518 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72e9a7b2-b482-4130-819a-95ac1106e218","Type":"ContainerStarted","Data":"479ab6973236d7632f68bb5b57d0888dc2c07de95ae8cfc1029ad9290416ff1d"} Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.367034 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.372563 4945 generic.go:334] "Generic (PLEG): container finished" podID="6cdd3739-5157-4344-acfc-8ff075bc91b8" containerID="f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a" exitCode=0 Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.372624 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" event={"ID":"6cdd3739-5157-4344-acfc-8ff075bc91b8","Type":"ContainerDied","Data":"f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a"} Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.372649 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" event={"ID":"6cdd3739-5157-4344-acfc-8ff075bc91b8","Type":"ContainerDied","Data":"d6b1b13bc047111f12b0f297c4cf72d283334e1b18d7c887a0304a2bba3713b4"} Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.372666 4945 scope.go:117] "RemoveContainer" containerID="f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.372787 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-khdrt" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.378856 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hr6hb" event={"ID":"aa503636-fff0-4195-9364-91c2f03e5c9b","Type":"ContainerDied","Data":"3ea3a894b50806fa707f0c629148a4127b12048b065b49fc31203eb5f82d2b6f"} Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.378910 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ea3a894b50806fa707f0c629148a4127b12048b065b49fc31203eb5f82d2b6f" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.378981 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hr6hb" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.423292 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.310882579 podStartE2EDuration="5.423261655s" podCreationTimestamp="2025-12-06 07:17:18 +0000 UTC" firstStartedPulling="2025-12-06 07:17:19.444226268 +0000 UTC m=+1452.899087312" lastFinishedPulling="2025-12-06 07:17:22.556605344 +0000 UTC m=+1456.011466388" observedRunningTime="2025-12-06 07:17:23.41060101 +0000 UTC m=+1456.865462054" watchObservedRunningTime="2025-12-06 07:17:23.423261655 +0000 UTC m=+1456.878122699" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.438798 4945 scope.go:117] "RemoveContainer" containerID="2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.440908 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6cdd3739-5157-4344-acfc-8ff075bc91b8" (UID: "6cdd3739-5157-4344-acfc-8ff075bc91b8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.452498 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.452526 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8rj5\" (UniqueName: \"kubernetes.io/projected/6cdd3739-5157-4344-acfc-8ff075bc91b8-kube-api-access-z8rj5\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.461172 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6cdd3739-5157-4344-acfc-8ff075bc91b8" (UID: "6cdd3739-5157-4344-acfc-8ff075bc91b8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.471473 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6cdd3739-5157-4344-acfc-8ff075bc91b8" (UID: "6cdd3739-5157-4344-acfc-8ff075bc91b8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.479537 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-config" (OuterVolumeSpecName: "config") pod "6cdd3739-5157-4344-acfc-8ff075bc91b8" (UID: "6cdd3739-5157-4344-acfc-8ff075bc91b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.513498 4945 scope.go:117] "RemoveContainer" containerID="f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.514695 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6cdd3739-5157-4344-acfc-8ff075bc91b8" (UID: "6cdd3739-5157-4344-acfc-8ff075bc91b8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:23 crc kubenswrapper[4945]: E1206 07:17:23.515186 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a\": container with ID starting with f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a not found: ID does not exist" containerID="f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.515235 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a"} err="failed to get container status \"f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a\": rpc error: code = NotFound desc = could not find container \"f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a\": container with ID starting with f2ad41437ea211a6b165270b221ae7116f79943d4f36eef9dfcc5a618cb3dd1a not found: ID does not exist" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.515265 4945 scope.go:117] "RemoveContainer" containerID="2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9" Dec 06 07:17:23 crc kubenswrapper[4945]: E1206 07:17:23.515888 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9\": container with ID starting with 2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9 not found: ID does not exist" containerID="2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.515931 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9"} err="failed to get container status \"2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9\": rpc error: code = NotFound desc = could not find container \"2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9\": container with ID starting with 2c547c81a02e63c8417c220f4d89a345b60737c70da275b2db4582cacded77d9 not found: ID does not exist" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.562801 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.563453 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.563544 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.563629 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cdd3739-5157-4344-acfc-8ff075bc91b8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.619398 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:17:23 crc kubenswrapper[4945]: E1206 07:17:23.619834 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cdd3739-5157-4344-acfc-8ff075bc91b8" containerName="init" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.619847 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cdd3739-5157-4344-acfc-8ff075bc91b8" containerName="init" Dec 06 07:17:23 crc kubenswrapper[4945]: E1206 07:17:23.619870 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cdd3739-5157-4344-acfc-8ff075bc91b8" containerName="dnsmasq-dns" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.619876 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cdd3739-5157-4344-acfc-8ff075bc91b8" containerName="dnsmasq-dns" Dec 06 07:17:23 crc kubenswrapper[4945]: E1206 07:17:23.619894 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa503636-fff0-4195-9364-91c2f03e5c9b" containerName="cinder-db-sync" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.619900 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa503636-fff0-4195-9364-91c2f03e5c9b" containerName="cinder-db-sync" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.620069 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cdd3739-5157-4344-acfc-8ff075bc91b8" containerName="dnsmasq-dns" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.620097 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa503636-fff0-4195-9364-91c2f03e5c9b" containerName="cinder-db-sync" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.621147 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.625175 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.625373 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.625478 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.625638 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-x9gc5" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.662470 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.721210 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-8svps"] Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.723367 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.745834 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-8svps"] Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.769553 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.769654 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.769731 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.769867 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqrlr\" (UniqueName: \"kubernetes.io/projected/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-kube-api-access-qqrlr\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.769933 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.770102 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-scripts\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.858899 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-khdrt"] Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876081 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-swift-storage-0\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876171 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876203 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876232 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876250 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-svc\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876297 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876322 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-config\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876351 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876368 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqrlr\" (UniqueName: \"kubernetes.io/projected/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-kube-api-access-qqrlr\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876390 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876433 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkpwn\" (UniqueName: \"kubernetes.io/projected/c6ac7f2c-e854-415e-9479-f808761d0a86-kube-api-access-pkpwn\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.876474 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-scripts\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.878444 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-khdrt"] Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.878564 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.888587 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.888746 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-scripts\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.899359 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.901611 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.903667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.907666 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.909184 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.910200 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqrlr\" (UniqueName: \"kubernetes.io/projected/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-kube-api-access-qqrlr\") pod \"cinder-scheduler-0\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.931895 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.978788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-config\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.978867 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.978990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkpwn\" (UniqueName: \"kubernetes.io/projected/c6ac7f2c-e854-415e-9479-f808761d0a86-kube-api-access-pkpwn\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.979099 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-swift-storage-0\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.979186 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-svc\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.979242 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.980352 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-swift-storage-0\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.981317 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-svc\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.981393 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-config\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.982247 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:23 crc kubenswrapper[4945]: I1206 07:17:23.982292 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.007035 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkpwn\" (UniqueName: \"kubernetes.io/projected/c6ac7f2c-e854-415e-9479-f808761d0a86-kube-api-access-pkpwn\") pod \"dnsmasq-dns-77d8c9c7-8svps\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.044937 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.075975 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.080655 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9qqg\" (UniqueName: \"kubernetes.io/projected/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-kube-api-access-r9qqg\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.080705 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.080827 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data-custom\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.080878 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.080902 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-scripts\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.080930 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-logs\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.080957 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.182563 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data-custom\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.182800 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.182831 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-scripts\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.182856 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-logs\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.182881 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.182981 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9qqg\" (UniqueName: \"kubernetes.io/projected/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-kube-api-access-r9qqg\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.183025 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.184774 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-logs\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.185330 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.198983 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.201892 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data-custom\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.206629 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-scripts\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.213416 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.215085 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9qqg\" (UniqueName: \"kubernetes.io/projected/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-kube-api-access-r9qqg\") pod \"cinder-api-0\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.362212 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.759507 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.765154 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-8svps"] Dec 06 07:17:24 crc kubenswrapper[4945]: W1206 07:17:24.772345 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0d381ab_0dbb_41f5_b1b7_b8c20d553cea.slice/crio-a9fd61ced63360b6ed56bad11512a0c7d19c1c8f31c92cd9054be748c894f1b5 WatchSource:0}: Error finding container a9fd61ced63360b6ed56bad11512a0c7d19c1c8f31c92cd9054be748c894f1b5: Status 404 returned error can't find the container with id a9fd61ced63360b6ed56bad11512a0c7d19c1c8f31c92cd9054be748c894f1b5 Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.952323 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:17:24 crc kubenswrapper[4945]: W1206 07:17:24.953368 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1acab70_a94a_4bfb_878e_a20c63b6c7f2.slice/crio-caf7c3b46300aa2dd0e198223a99a610ef7571d05e539054a5d7f5d2af41eb1b WatchSource:0}: Error finding container caf7c3b46300aa2dd0e198223a99a610ef7571d05e539054a5d7f5d2af41eb1b: Status 404 returned error can't find the container with id caf7c3b46300aa2dd0e198223a99a610ef7571d05e539054a5d7f5d2af41eb1b Dec 06 07:17:24 crc kubenswrapper[4945]: I1206 07:17:24.967729 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cdd3739-5157-4344-acfc-8ff075bc91b8" path="/var/lib/kubelet/pods/6cdd3739-5157-4344-acfc-8ff075bc91b8/volumes" Dec 06 07:17:25 crc kubenswrapper[4945]: I1206 07:17:25.135205 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:25 crc kubenswrapper[4945]: I1206 07:17:25.245804 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:25 crc kubenswrapper[4945]: I1206 07:17:25.435133 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea","Type":"ContainerStarted","Data":"a9fd61ced63360b6ed56bad11512a0c7d19c1c8f31c92cd9054be748c894f1b5"} Dec 06 07:17:25 crc kubenswrapper[4945]: I1206 07:17:25.437798 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a1acab70-a94a-4bfb-878e-a20c63b6c7f2","Type":"ContainerStarted","Data":"caf7c3b46300aa2dd0e198223a99a610ef7571d05e539054a5d7f5d2af41eb1b"} Dec 06 07:17:25 crc kubenswrapper[4945]: I1206 07:17:25.441705 4945 generic.go:334] "Generic (PLEG): container finished" podID="c6ac7f2c-e854-415e-9479-f808761d0a86" containerID="1484408d2bfa10e900e12b9b1944796a40e33648452c091db90f31342c12ad9f" exitCode=0 Dec 06 07:17:25 crc kubenswrapper[4945]: I1206 07:17:25.443724 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" event={"ID":"c6ac7f2c-e854-415e-9479-f808761d0a86","Type":"ContainerDied","Data":"1484408d2bfa10e900e12b9b1944796a40e33648452c091db90f31342c12ad9f"} Dec 06 07:17:25 crc kubenswrapper[4945]: I1206 07:17:25.444449 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" event={"ID":"c6ac7f2c-e854-415e-9479-f808761d0a86","Type":"ContainerStarted","Data":"ffa08229dd0758d42105d6513d2c2a445816fc8a39657db60371062ecf97b6f8"} Dec 06 07:17:26 crc kubenswrapper[4945]: I1206 07:17:26.472529 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a1acab70-a94a-4bfb-878e-a20c63b6c7f2","Type":"ContainerStarted","Data":"a2852e09c45fa23e140608ec02c569334b53cf80a99edae75b75522c7d48f454"} Dec 06 07:17:26 crc kubenswrapper[4945]: I1206 07:17:26.489727 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" event={"ID":"c6ac7f2c-e854-415e-9479-f808761d0a86","Type":"ContainerStarted","Data":"97a7cb8a000b2b0ca98052a7ec3df67c8c15d0de75ce40f5229d9b011bb3c1b0"} Dec 06 07:17:26 crc kubenswrapper[4945]: I1206 07:17:26.491091 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:26 crc kubenswrapper[4945]: I1206 07:17:26.522592 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" podStartSLOduration=3.522569024 podStartE2EDuration="3.522569024s" podCreationTimestamp="2025-12-06 07:17:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:26.518665578 +0000 UTC m=+1459.973526622" watchObservedRunningTime="2025-12-06 07:17:26.522569024 +0000 UTC m=+1459.977430068" Dec 06 07:17:26 crc kubenswrapper[4945]: I1206 07:17:26.641753 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:17:27 crc kubenswrapper[4945]: I1206 07:17:27.518002 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea","Type":"ContainerStarted","Data":"bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41"} Dec 06 07:17:27 crc kubenswrapper[4945]: I1206 07:17:27.527495 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a1acab70-a94a-4bfb-878e-a20c63b6c7f2","Type":"ContainerStarted","Data":"b5d25d4b24441a520223e8b318d2e0dca92158e2459d09ca919b3aafc1b0d487"} Dec 06 07:17:27 crc kubenswrapper[4945]: I1206 07:17:27.527719 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" containerName="cinder-api-log" containerID="cri-o://a2852e09c45fa23e140608ec02c569334b53cf80a99edae75b75522c7d48f454" gracePeriod=30 Dec 06 07:17:27 crc kubenswrapper[4945]: I1206 07:17:27.528323 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" containerName="cinder-api" containerID="cri-o://b5d25d4b24441a520223e8b318d2e0dca92158e2459d09ca919b3aafc1b0d487" gracePeriod=30 Dec 06 07:17:27 crc kubenswrapper[4945]: I1206 07:17:27.551636 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.551614063 podStartE2EDuration="4.551614063s" podCreationTimestamp="2025-12-06 07:17:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:27.550379089 +0000 UTC m=+1461.005240143" watchObservedRunningTime="2025-12-06 07:17:27.551614063 +0000 UTC m=+1461.006475107" Dec 06 07:17:28 crc kubenswrapper[4945]: I1206 07:17:28.544218 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea","Type":"ContainerStarted","Data":"b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe"} Dec 06 07:17:28 crc kubenswrapper[4945]: I1206 07:17:28.565121 4945 generic.go:334] "Generic (PLEG): container finished" podID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" containerID="a2852e09c45fa23e140608ec02c569334b53cf80a99edae75b75522c7d48f454" exitCode=143 Dec 06 07:17:28 crc kubenswrapper[4945]: I1206 07:17:28.565330 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a1acab70-a94a-4bfb-878e-a20c63b6c7f2","Type":"ContainerDied","Data":"a2852e09c45fa23e140608ec02c569334b53cf80a99edae75b75522c7d48f454"} Dec 06 07:17:28 crc kubenswrapper[4945]: I1206 07:17:28.577782 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.642922338 podStartE2EDuration="5.577763154s" podCreationTimestamp="2025-12-06 07:17:23 +0000 UTC" firstStartedPulling="2025-12-06 07:17:24.778597002 +0000 UTC m=+1458.233458046" lastFinishedPulling="2025-12-06 07:17:25.713437818 +0000 UTC m=+1459.168298862" observedRunningTime="2025-12-06 07:17:28.57138965 +0000 UTC m=+1462.026250704" watchObservedRunningTime="2025-12-06 07:17:28.577763154 +0000 UTC m=+1462.032624198" Dec 06 07:17:28 crc kubenswrapper[4945]: I1206 07:17:28.774467 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:28 crc kubenswrapper[4945]: I1206 07:17:28.984550 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:17:29 crc kubenswrapper[4945]: I1206 07:17:29.045522 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 07:17:29 crc kubenswrapper[4945]: I1206 07:17:29.070213 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5899f79f4b-lsl4d"] Dec 06 07:17:29 crc kubenswrapper[4945]: I1206 07:17:29.070822 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5899f79f4b-lsl4d" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api-log" containerID="cri-o://023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822" gracePeriod=30 Dec 06 07:17:29 crc kubenswrapper[4945]: I1206 07:17:29.071382 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5899f79f4b-lsl4d" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api" containerID="cri-o://189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a" gracePeriod=30 Dec 06 07:17:29 crc kubenswrapper[4945]: I1206 07:17:29.081655 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5899f79f4b-lsl4d" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": EOF" Dec 06 07:17:29 crc kubenswrapper[4945]: I1206 07:17:29.363403 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 07:17:29 crc kubenswrapper[4945]: I1206 07:17:29.579581 4945 generic.go:334] "Generic (PLEG): container finished" podID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerID="023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822" exitCode=143 Dec 06 07:17:29 crc kubenswrapper[4945]: I1206 07:17:29.580692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5899f79f4b-lsl4d" event={"ID":"9d840c09-0868-449e-9699-9b84aa8e0fce","Type":"ContainerDied","Data":"023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822"} Dec 06 07:17:31 crc kubenswrapper[4945]: I1206 07:17:31.997080 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:32 crc kubenswrapper[4945]: I1206 07:17:32.081660 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:17:32 crc kubenswrapper[4945]: I1206 07:17:32.393717 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:32 crc kubenswrapper[4945]: I1206 07:17:32.898875 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:17:33 crc kubenswrapper[4945]: I1206 07:17:33.625938 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5899f79f4b-lsl4d" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:49974->10.217.0.156:9311: read: connection reset by peer" Dec 06 07:17:33 crc kubenswrapper[4945]: I1206 07:17:33.625999 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5899f79f4b-lsl4d" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:49976->10.217.0.156:9311: read: connection reset by peer" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.079479 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.120349 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.155637 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-7kgt2"] Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.162572 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" podUID="b75c5b75-ef89-4898-aca5-7c3bd94cda90" containerName="dnsmasq-dns" containerID="cri-o://1f2f55eba30b22190ba637ec820a6dc48392790ff0937bf474cf5b424bdb9e78" gracePeriod=10 Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.247071 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data\") pod \"9d840c09-0868-449e-9699-9b84aa8e0fce\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.247156 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data-custom\") pod \"9d840c09-0868-449e-9699-9b84aa8e0fce\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.247217 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d840c09-0868-449e-9699-9b84aa8e0fce-logs\") pod \"9d840c09-0868-449e-9699-9b84aa8e0fce\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.247371 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p29lc\" (UniqueName: \"kubernetes.io/projected/9d840c09-0868-449e-9699-9b84aa8e0fce-kube-api-access-p29lc\") pod \"9d840c09-0868-449e-9699-9b84aa8e0fce\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.247431 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-combined-ca-bundle\") pod \"9d840c09-0868-449e-9699-9b84aa8e0fce\" (UID: \"9d840c09-0868-449e-9699-9b84aa8e0fce\") " Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.250056 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d840c09-0868-449e-9699-9b84aa8e0fce-logs" (OuterVolumeSpecName: "logs") pod "9d840c09-0868-449e-9699-9b84aa8e0fce" (UID: "9d840c09-0868-449e-9699-9b84aa8e0fce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.289620 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9d840c09-0868-449e-9699-9b84aa8e0fce" (UID: "9d840c09-0868-449e-9699-9b84aa8e0fce"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.322491 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d840c09-0868-449e-9699-9b84aa8e0fce-kube-api-access-p29lc" (OuterVolumeSpecName: "kube-api-access-p29lc") pod "9d840c09-0868-449e-9699-9b84aa8e0fce" (UID: "9d840c09-0868-449e-9699-9b84aa8e0fce"). InnerVolumeSpecName "kube-api-access-p29lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.350348 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.350680 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d840c09-0868-449e-9699-9b84aa8e0fce-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.350701 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p29lc\" (UniqueName: \"kubernetes.io/projected/9d840c09-0868-449e-9699-9b84aa8e0fce-kube-api-access-p29lc\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.420551 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d840c09-0868-449e-9699-9b84aa8e0fce" (UID: "9d840c09-0868-449e-9699-9b84aa8e0fce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.435846 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data" (OuterVolumeSpecName: "config-data") pod "9d840c09-0868-449e-9699-9b84aa8e0fce" (UID: "9d840c09-0868-449e-9699-9b84aa8e0fce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.453739 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.453774 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d840c09-0868-449e-9699-9b84aa8e0fce-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.547274 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.590978 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.634240 4945 generic.go:334] "Generic (PLEG): container finished" podID="b75c5b75-ef89-4898-aca5-7c3bd94cda90" containerID="1f2f55eba30b22190ba637ec820a6dc48392790ff0937bf474cf5b424bdb9e78" exitCode=0 Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.634325 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" event={"ID":"b75c5b75-ef89-4898-aca5-7c3bd94cda90","Type":"ContainerDied","Data":"1f2f55eba30b22190ba637ec820a6dc48392790ff0937bf474cf5b424bdb9e78"} Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.638535 4945 generic.go:334] "Generic (PLEG): container finished" podID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerID="189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a" exitCode=0 Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.638581 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5899f79f4b-lsl4d" event={"ID":"9d840c09-0868-449e-9699-9b84aa8e0fce","Type":"ContainerDied","Data":"189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a"} Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.638596 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5899f79f4b-lsl4d" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.638629 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5899f79f4b-lsl4d" event={"ID":"9d840c09-0868-449e-9699-9b84aa8e0fce","Type":"ContainerDied","Data":"800d7c1faa962d8af183da7c56ddcb721fb6ffd71ca900156565b59649d7d43d"} Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.638651 4945 scope.go:117] "RemoveContainer" containerID="189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.638945 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" containerName="cinder-scheduler" containerID="cri-o://bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41" gracePeriod=30 Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.638996 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" containerName="probe" containerID="cri-o://b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe" gracePeriod=30 Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.689561 4945 scope.go:117] "RemoveContainer" containerID="023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.701342 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5899f79f4b-lsl4d"] Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.711905 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5899f79f4b-lsl4d"] Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.754407 4945 scope.go:117] "RemoveContainer" containerID="189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a" Dec 06 07:17:34 crc kubenswrapper[4945]: E1206 07:17:34.755071 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a\": container with ID starting with 189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a not found: ID does not exist" containerID="189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.755106 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a"} err="failed to get container status \"189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a\": rpc error: code = NotFound desc = could not find container \"189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a\": container with ID starting with 189ce8a10356f80bad157575b6de3a38d39f8b3896c5b23234f2105241b0e68a not found: ID does not exist" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.755129 4945 scope.go:117] "RemoveContainer" containerID="023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822" Dec 06 07:17:34 crc kubenswrapper[4945]: E1206 07:17:34.755784 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822\": container with ID starting with 023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822 not found: ID does not exist" containerID="023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.755846 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822"} err="failed to get container status \"023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822\": rpc error: code = NotFound desc = could not find container \"023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822\": container with ID starting with 023b0e5157445be611a145a79b9dda4a382683e282ffcca67dff015a4b9ed822 not found: ID does not exist" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.977672 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" path="/var/lib/kubelet/pods/9d840c09-0868-449e-9699-9b84aa8e0fce/volumes" Dec 06 07:17:34 crc kubenswrapper[4945]: I1206 07:17:34.978510 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.059294 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-577494887b-gjx4w"] Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.059798 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-577494887b-gjx4w" podUID="fc779c1d-c266-43b3-95b0-8a799da93e7c" containerName="neutron-api" containerID="cri-o://a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c" gracePeriod=30 Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.059947 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-577494887b-gjx4w" podUID="fc779c1d-c266-43b3-95b0-8a799da93e7c" containerName="neutron-httpd" containerID="cri-o://bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48" gracePeriod=30 Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.289602 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.386397 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-svc\") pod \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.386473 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-sb\") pod \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.386625 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-swift-storage-0\") pod \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.386683 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmvjn\" (UniqueName: \"kubernetes.io/projected/b75c5b75-ef89-4898-aca5-7c3bd94cda90-kube-api-access-xmvjn\") pod \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.386718 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-nb\") pod \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.386867 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-config\") pod \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\" (UID: \"b75c5b75-ef89-4898-aca5-7c3bd94cda90\") " Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.393844 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b75c5b75-ef89-4898-aca5-7c3bd94cda90-kube-api-access-xmvjn" (OuterVolumeSpecName: "kube-api-access-xmvjn") pod "b75c5b75-ef89-4898-aca5-7c3bd94cda90" (UID: "b75c5b75-ef89-4898-aca5-7c3bd94cda90"). InnerVolumeSpecName "kube-api-access-xmvjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.482002 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b75c5b75-ef89-4898-aca5-7c3bd94cda90" (UID: "b75c5b75-ef89-4898-aca5-7c3bd94cda90"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.490846 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.490886 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmvjn\" (UniqueName: \"kubernetes.io/projected/b75c5b75-ef89-4898-aca5-7c3bd94cda90-kube-api-access-xmvjn\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.491814 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b75c5b75-ef89-4898-aca5-7c3bd94cda90" (UID: "b75c5b75-ef89-4898-aca5-7c3bd94cda90"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.496686 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b75c5b75-ef89-4898-aca5-7c3bd94cda90" (UID: "b75c5b75-ef89-4898-aca5-7c3bd94cda90"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.503581 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b75c5b75-ef89-4898-aca5-7c3bd94cda90" (UID: "b75c5b75-ef89-4898-aca5-7c3bd94cda90"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.507777 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-config" (OuterVolumeSpecName: "config") pod "b75c5b75-ef89-4898-aca5-7c3bd94cda90" (UID: "b75c5b75-ef89-4898-aca5-7c3bd94cda90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.592643 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.592890 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.592949 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.593007 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b75c5b75-ef89-4898-aca5-7c3bd94cda90-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.650325 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-577494887b-gjx4w" event={"ID":"fc779c1d-c266-43b3-95b0-8a799da93e7c","Type":"ContainerDied","Data":"bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48"} Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.650258 4945 generic.go:334] "Generic (PLEG): container finished" podID="fc779c1d-c266-43b3-95b0-8a799da93e7c" containerID="bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48" exitCode=0 Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.653486 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" event={"ID":"b75c5b75-ef89-4898-aca5-7c3bd94cda90","Type":"ContainerDied","Data":"25e52c74a72794d95143f897ac67d7b895460dcd93696d15d4b17da9b6ee85f2"} Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.653639 4945 scope.go:117] "RemoveContainer" containerID="1f2f55eba30b22190ba637ec820a6dc48392790ff0937bf474cf5b424bdb9e78" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.653671 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-7kgt2" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.694600 4945 scope.go:117] "RemoveContainer" containerID="14844dfa9e120ac477947241791eb4b53a707c2e73c9bd37e9218941aae58d93" Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.697792 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-7kgt2"] Dec 06 07:17:35 crc kubenswrapper[4945]: I1206 07:17:35.705634 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-7kgt2"] Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.671620 4945 generic.go:334] "Generic (PLEG): container finished" podID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" containerID="b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe" exitCode=0 Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.671697 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea","Type":"ContainerDied","Data":"b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe"} Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.802066 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 06 07:17:36 crc kubenswrapper[4945]: E1206 07:17:36.802822 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b75c5b75-ef89-4898-aca5-7c3bd94cda90" containerName="dnsmasq-dns" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.802842 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b75c5b75-ef89-4898-aca5-7c3bd94cda90" containerName="dnsmasq-dns" Dec 06 07:17:36 crc kubenswrapper[4945]: E1206 07:17:36.802868 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api-log" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.802875 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api-log" Dec 06 07:17:36 crc kubenswrapper[4945]: E1206 07:17:36.802901 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b75c5b75-ef89-4898-aca5-7c3bd94cda90" containerName="init" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.802908 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b75c5b75-ef89-4898-aca5-7c3bd94cda90" containerName="init" Dec 06 07:17:36 crc kubenswrapper[4945]: E1206 07:17:36.802926 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.802934 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.803108 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api-log" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.803119 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d840c09-0868-449e-9699-9b84aa8e0fce" containerName="barbican-api" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.803133 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b75c5b75-ef89-4898-aca5-7c3bd94cda90" containerName="dnsmasq-dns" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.803972 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.806191 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.806391 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.806622 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-k4c4q" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.815341 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.886800 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.916484 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.916544 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.916844 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config-secret\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.916988 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjvvw\" (UniqueName: \"kubernetes.io/projected/7e90dc2b-f137-411b-87e4-b46f36d3034e-kube-api-access-hjvvw\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:36 crc kubenswrapper[4945]: I1206 07:17:36.964606 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b75c5b75-ef89-4898-aca5-7c3bd94cda90" path="/var/lib/kubelet/pods/b75c5b75-ef89-4898-aca5-7c3bd94cda90/volumes" Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.018661 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.018731 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.018782 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config-secret\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.018835 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjvvw\" (UniqueName: \"kubernetes.io/projected/7e90dc2b-f137-411b-87e4-b46f36d3034e-kube-api-access-hjvvw\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.019780 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.027303 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.028786 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config-secret\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.049370 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjvvw\" (UniqueName: \"kubernetes.io/projected/7e90dc2b-f137-411b-87e4-b46f36d3034e-kube-api-access-hjvvw\") pod \"openstackclient\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " pod="openstack/openstackclient" Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.121617 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.600580 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 07:17:37 crc kubenswrapper[4945]: W1206 07:17:37.605741 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e90dc2b_f137_411b_87e4_b46f36d3034e.slice/crio-1e783bd38c3a61bdbfa595382e44b07597d899b57aae7fabfd62fd62abb2cc0d WatchSource:0}: Error finding container 1e783bd38c3a61bdbfa595382e44b07597d899b57aae7fabfd62fd62abb2cc0d: Status 404 returned error can't find the container with id 1e783bd38c3a61bdbfa595382e44b07597d899b57aae7fabfd62fd62abb2cc0d Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.609340 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:17:37 crc kubenswrapper[4945]: I1206 07:17:37.685186 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"7e90dc2b-f137-411b-87e4-b46f36d3034e","Type":"ContainerStarted","Data":"1e783bd38c3a61bdbfa595382e44b07597d899b57aae7fabfd62fd62abb2cc0d"} Dec 06 07:17:38 crc kubenswrapper[4945]: I1206 07:17:38.797562 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:17:38 crc kubenswrapper[4945]: I1206 07:17:38.798132 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:17:38 crc kubenswrapper[4945]: I1206 07:17:38.798198 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:17:38 crc kubenswrapper[4945]: I1206 07:17:38.799771 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6db8402445675a35c1cfea5f14ca674ed295b8ef4463d76bc429b4cab4e4331a"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:17:38 crc kubenswrapper[4945]: I1206 07:17:38.799871 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://6db8402445675a35c1cfea5f14ca674ed295b8ef4463d76bc429b4cab4e4331a" gracePeriod=600 Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.555442 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.610427 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data-custom\") pod \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.610473 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-combined-ca-bundle\") pod \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.610510 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-scripts\") pod \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.610581 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data\") pod \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.610657 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-etc-machine-id\") pod \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.610685 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqrlr\" (UniqueName: \"kubernetes.io/projected/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-kube-api-access-qqrlr\") pod \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\" (UID: \"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea\") " Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.612049 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" (UID: "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.625516 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" (UID: "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.625558 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-kube-api-access-qqrlr" (OuterVolumeSpecName: "kube-api-access-qqrlr") pod "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" (UID: "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea"). InnerVolumeSpecName "kube-api-access-qqrlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.625566 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-scripts" (OuterVolumeSpecName: "scripts") pod "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" (UID: "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.712921 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.712949 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.712958 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.712969 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqrlr\" (UniqueName: \"kubernetes.io/projected/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-kube-api-access-qqrlr\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.719410 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" (UID: "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.724646 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="6db8402445675a35c1cfea5f14ca674ed295b8ef4463d76bc429b4cab4e4331a" exitCode=0 Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.724769 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"6db8402445675a35c1cfea5f14ca674ed295b8ef4463d76bc429b4cab4e4331a"} Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.724813 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed"} Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.724833 4945 scope.go:117] "RemoveContainer" containerID="fd69474a2e644df9791d0d3744c09cd9214dcea77b7b07f5d390844d87230bc1" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.734428 4945 generic.go:334] "Generic (PLEG): container finished" podID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" containerID="bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41" exitCode=0 Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.734472 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea","Type":"ContainerDied","Data":"bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41"} Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.734500 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a0d381ab-0dbb-41f5-b1b7-b8c20d553cea","Type":"ContainerDied","Data":"a9fd61ced63360b6ed56bad11512a0c7d19c1c8f31c92cd9054be748c894f1b5"} Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.734668 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.766429 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data" (OuterVolumeSpecName: "config-data") pod "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" (UID: "a0d381ab-0dbb-41f5-b1b7-b8c20d553cea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.769323 4945 scope.go:117] "RemoveContainer" containerID="b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.804783 4945 scope.go:117] "RemoveContainer" containerID="bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.816776 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.816820 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.878466 4945 scope.go:117] "RemoveContainer" containerID="b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe" Dec 06 07:17:39 crc kubenswrapper[4945]: E1206 07:17:39.881444 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe\": container with ID starting with b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe not found: ID does not exist" containerID="b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.881485 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe"} err="failed to get container status \"b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe\": rpc error: code = NotFound desc = could not find container \"b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe\": container with ID starting with b51a4209eff0515c5e750894268f82c6e1c404405add6038a2df016fe79f6efe not found: ID does not exist" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.881510 4945 scope.go:117] "RemoveContainer" containerID="bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41" Dec 06 07:17:39 crc kubenswrapper[4945]: E1206 07:17:39.884058 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41\": container with ID starting with bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41 not found: ID does not exist" containerID="bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41" Dec 06 07:17:39 crc kubenswrapper[4945]: I1206 07:17:39.884109 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41"} err="failed to get container status \"bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41\": rpc error: code = NotFound desc = could not find container \"bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41\": container with ID starting with bd7da95f4513e78fd4e9d13a39bbbb74003321ae51194096eeea4905ad2b8d41 not found: ID does not exist" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.097331 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.120181 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.144335 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:17:40 crc kubenswrapper[4945]: E1206 07:17:40.144895 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" containerName="probe" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.144921 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" containerName="probe" Dec 06 07:17:40 crc kubenswrapper[4945]: E1206 07:17:40.144970 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" containerName="cinder-scheduler" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.144979 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" containerName="cinder-scheduler" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.145208 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" containerName="probe" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.145235 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" containerName="cinder-scheduler" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.146631 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.158143 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.162088 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.226159 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.226571 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d74c1f-98d7-40ac-a1e8-233604971bc9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.226620 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.226734 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-scripts\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.226768 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkmkq\" (UniqueName: \"kubernetes.io/projected/b6d74c1f-98d7-40ac-a1e8-233604971bc9-kube-api-access-mkmkq\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.226792 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.329636 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.329893 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d74c1f-98d7-40ac-a1e8-233604971bc9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.330008 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.330034 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d74c1f-98d7-40ac-a1e8-233604971bc9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.330239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-scripts\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.330350 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkmkq\" (UniqueName: \"kubernetes.io/projected/b6d74c1f-98d7-40ac-a1e8-233604971bc9-kube-api-access-mkmkq\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.330659 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.338751 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.342759 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-scripts\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.345065 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.363166 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.365896 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkmkq\" (UniqueName: \"kubernetes.io/projected/b6d74c1f-98d7-40ac-a1e8-233604971bc9-kube-api-access-mkmkq\") pod \"cinder-scheduler-0\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.495884 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.610096 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.637264 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m99xw\" (UniqueName: \"kubernetes.io/projected/fc779c1d-c266-43b3-95b0-8a799da93e7c-kube-api-access-m99xw\") pod \"fc779c1d-c266-43b3-95b0-8a799da93e7c\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.637935 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-httpd-config\") pod \"fc779c1d-c266-43b3-95b0-8a799da93e7c\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.638179 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-combined-ca-bundle\") pod \"fc779c1d-c266-43b3-95b0-8a799da93e7c\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.638228 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-config\") pod \"fc779c1d-c266-43b3-95b0-8a799da93e7c\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.638307 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-ovndb-tls-certs\") pod \"fc779c1d-c266-43b3-95b0-8a799da93e7c\" (UID: \"fc779c1d-c266-43b3-95b0-8a799da93e7c\") " Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.652675 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "fc779c1d-c266-43b3-95b0-8a799da93e7c" (UID: "fc779c1d-c266-43b3-95b0-8a799da93e7c"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.652926 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc779c1d-c266-43b3-95b0-8a799da93e7c-kube-api-access-m99xw" (OuterVolumeSpecName: "kube-api-access-m99xw") pod "fc779c1d-c266-43b3-95b0-8a799da93e7c" (UID: "fc779c1d-c266-43b3-95b0-8a799da93e7c"). InnerVolumeSpecName "kube-api-access-m99xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.706704 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc779c1d-c266-43b3-95b0-8a799da93e7c" (UID: "fc779c1d-c266-43b3-95b0-8a799da93e7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.712872 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-config" (OuterVolumeSpecName: "config") pod "fc779c1d-c266-43b3-95b0-8a799da93e7c" (UID: "fc779c1d-c266-43b3-95b0-8a799da93e7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.719986 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "fc779c1d-c266-43b3-95b0-8a799da93e7c" (UID: "fc779c1d-c266-43b3-95b0-8a799da93e7c"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.741906 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.741940 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.741953 4945 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.741996 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m99xw\" (UniqueName: \"kubernetes.io/projected/fc779c1d-c266-43b3-95b0-8a799da93e7c-kube-api-access-m99xw\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.742013 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fc779c1d-c266-43b3-95b0-8a799da93e7c-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.784598 4945 generic.go:334] "Generic (PLEG): container finished" podID="fc779c1d-c266-43b3-95b0-8a799da93e7c" containerID="a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c" exitCode=0 Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.784648 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-577494887b-gjx4w" event={"ID":"fc779c1d-c266-43b3-95b0-8a799da93e7c","Type":"ContainerDied","Data":"a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c"} Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.784676 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-577494887b-gjx4w" event={"ID":"fc779c1d-c266-43b3-95b0-8a799da93e7c","Type":"ContainerDied","Data":"fe3414acc3d351c832571d8b08ef783dbd5a90259e360add7975c6b8fb5156d7"} Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.784694 4945 scope.go:117] "RemoveContainer" containerID="bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.784969 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-577494887b-gjx4w" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.842992 4945 scope.go:117] "RemoveContainer" containerID="a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.860488 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-577494887b-gjx4w"] Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.881939 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-577494887b-gjx4w"] Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.892342 4945 scope.go:117] "RemoveContainer" containerID="bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48" Dec 06 07:17:40 crc kubenswrapper[4945]: E1206 07:17:40.894536 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48\": container with ID starting with bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48 not found: ID does not exist" containerID="bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.894578 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48"} err="failed to get container status \"bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48\": rpc error: code = NotFound desc = could not find container \"bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48\": container with ID starting with bfb738143d73e9693584bb18160ed30ad3ab88e0fc80adf14692fea5ea0f4b48 not found: ID does not exist" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.894612 4945 scope.go:117] "RemoveContainer" containerID="a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c" Dec 06 07:17:40 crc kubenswrapper[4945]: E1206 07:17:40.897242 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c\": container with ID starting with a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c not found: ID does not exist" containerID="a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.897331 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c"} err="failed to get container status \"a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c\": rpc error: code = NotFound desc = could not find container \"a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c\": container with ID starting with a7077d23015fbcd8375293ae5e976b7e5edb5da7da5b47cc9eb2e8900157191c not found: ID does not exist" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.967131 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0d381ab-0dbb-41f5-b1b7-b8c20d553cea" path="/var/lib/kubelet/pods/a0d381ab-0dbb-41f5-b1b7-b8c20d553cea/volumes" Dec 06 07:17:40 crc kubenswrapper[4945]: I1206 07:17:40.968252 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc779c1d-c266-43b3-95b0-8a799da93e7c" path="/var/lib/kubelet/pods/fc779c1d-c266-43b3-95b0-8a799da93e7c/volumes" Dec 06 07:17:41 crc kubenswrapper[4945]: W1206 07:17:41.015802 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6d74c1f_98d7_40ac_a1e8_233604971bc9.slice/crio-499f462096d17366e60f722095d02f3e598561067542dcee4e671d295dc8df3f WatchSource:0}: Error finding container 499f462096d17366e60f722095d02f3e598561067542dcee4e671d295dc8df3f: Status 404 returned error can't find the container with id 499f462096d17366e60f722095d02f3e598561067542dcee4e671d295dc8df3f Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.031011 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.757537 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.758066 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="ceilometer-central-agent" containerID="cri-o://2139e668db9a92f5f7706eff1e6c504a57c97c94612cadb38beac36f2fa9ed10" gracePeriod=30 Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.758504 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="ceilometer-notification-agent" containerID="cri-o://3ab75bae5874506e9ae523bd85a3d001e584020c211168290d995e46abd09b61" gracePeriod=30 Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.758494 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="sg-core" containerID="cri-o://df489374460e08de26f48143dd596d770a90fe2c61e81c40c9410a4bcd4d5df6" gracePeriod=30 Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.758806 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="proxy-httpd" containerID="cri-o://479ab6973236d7632f68bb5b57d0888dc2c07de95ae8cfc1029ad9290416ff1d" gracePeriod=30 Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.798830 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b6d74c1f-98d7-40ac-a1e8-233604971bc9","Type":"ContainerStarted","Data":"499f462096d17366e60f722095d02f3e598561067542dcee4e671d295dc8df3f"} Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.799010 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": EOF" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.903662 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6ccd85df7-25dfh"] Dec 06 07:17:41 crc kubenswrapper[4945]: E1206 07:17:41.904347 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc779c1d-c266-43b3-95b0-8a799da93e7c" containerName="neutron-api" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.904366 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc779c1d-c266-43b3-95b0-8a799da93e7c" containerName="neutron-api" Dec 06 07:17:41 crc kubenswrapper[4945]: E1206 07:17:41.904400 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc779c1d-c266-43b3-95b0-8a799da93e7c" containerName="neutron-httpd" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.904408 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc779c1d-c266-43b3-95b0-8a799da93e7c" containerName="neutron-httpd" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.905054 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc779c1d-c266-43b3-95b0-8a799da93e7c" containerName="neutron-api" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.905077 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc779c1d-c266-43b3-95b0-8a799da93e7c" containerName="neutron-httpd" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.906526 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.910395 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.910700 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.910982 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.965378 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6ccd85df7-25dfh"] Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.993705 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-run-httpd\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.993771 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-combined-ca-bundle\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.993809 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-public-tls-certs\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.993850 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-log-httpd\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.993878 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-etc-swift\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.993933 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnr6p\" (UniqueName: \"kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-kube-api-access-jnr6p\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.993982 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-internal-tls-certs\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:41 crc kubenswrapper[4945]: I1206 07:17:41.994018 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-config-data\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.100919 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-run-httpd\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.100985 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-combined-ca-bundle\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.101017 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-public-tls-certs\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.101057 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-log-httpd\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.101085 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-etc-swift\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.101164 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnr6p\" (UniqueName: \"kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-kube-api-access-jnr6p\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.101217 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-internal-tls-certs\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.101255 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-config-data\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.103855 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-log-httpd\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.104127 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-run-httpd\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.112555 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-etc-swift\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.112565 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-config-data\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.120610 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-public-tls-certs\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.122155 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-combined-ca-bundle\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.133474 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-internal-tls-certs\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.145086 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnr6p\" (UniqueName: \"kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-kube-api-access-jnr6p\") pod \"swift-proxy-6ccd85df7-25dfh\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.304804 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.822908 4945 generic.go:334] "Generic (PLEG): container finished" podID="72e9a7b2-b482-4130-819a-95ac1106e218" containerID="479ab6973236d7632f68bb5b57d0888dc2c07de95ae8cfc1029ad9290416ff1d" exitCode=0 Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.823230 4945 generic.go:334] "Generic (PLEG): container finished" podID="72e9a7b2-b482-4130-819a-95ac1106e218" containerID="df489374460e08de26f48143dd596d770a90fe2c61e81c40c9410a4bcd4d5df6" exitCode=2 Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.822971 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72e9a7b2-b482-4130-819a-95ac1106e218","Type":"ContainerDied","Data":"479ab6973236d7632f68bb5b57d0888dc2c07de95ae8cfc1029ad9290416ff1d"} Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.823243 4945 generic.go:334] "Generic (PLEG): container finished" podID="72e9a7b2-b482-4130-819a-95ac1106e218" containerID="2139e668db9a92f5f7706eff1e6c504a57c97c94612cadb38beac36f2fa9ed10" exitCode=0 Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.823272 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72e9a7b2-b482-4130-819a-95ac1106e218","Type":"ContainerDied","Data":"df489374460e08de26f48143dd596d770a90fe2c61e81c40c9410a4bcd4d5df6"} Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.823344 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72e9a7b2-b482-4130-819a-95ac1106e218","Type":"ContainerDied","Data":"2139e668db9a92f5f7706eff1e6c504a57c97c94612cadb38beac36f2fa9ed10"} Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.825466 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b6d74c1f-98d7-40ac-a1e8-233604971bc9","Type":"ContainerStarted","Data":"14269f510b2f262408ff083c20bf00b7f68b48a2e3c497d57f6d2ea6e8cc4fcb"} Dec 06 07:17:42 crc kubenswrapper[4945]: I1206 07:17:42.986402 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6ccd85df7-25dfh"] Dec 06 07:17:43 crc kubenswrapper[4945]: I1206 07:17:43.844147 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6ccd85df7-25dfh" event={"ID":"c64f8a71-9ee5-4145-b635-bd282dabd781","Type":"ContainerStarted","Data":"29a489e1a3fc4b892c014063d26ad8a5a7da41074019dcb533e94027b8e256f2"} Dec 06 07:17:43 crc kubenswrapper[4945]: I1206 07:17:43.844540 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6ccd85df7-25dfh" event={"ID":"c64f8a71-9ee5-4145-b635-bd282dabd781","Type":"ContainerStarted","Data":"6afab1fa21766e8cbfdedb4c1868e051427546f3e9b764dfcadfaac9ac9f3aa3"} Dec 06 07:17:43 crc kubenswrapper[4945]: I1206 07:17:43.844553 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6ccd85df7-25dfh" event={"ID":"c64f8a71-9ee5-4145-b635-bd282dabd781","Type":"ContainerStarted","Data":"db08e54665d085bc316da8216960b9336d36d7829250120b04981c2d5544f1bd"} Dec 06 07:17:43 crc kubenswrapper[4945]: I1206 07:17:43.845087 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:43 crc kubenswrapper[4945]: I1206 07:17:43.845180 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:43 crc kubenswrapper[4945]: I1206 07:17:43.854304 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b6d74c1f-98d7-40ac-a1e8-233604971bc9","Type":"ContainerStarted","Data":"1041e9cc315c9c6b1e1b5a4a949b745e06f5a68d6531286428930d1c1eb3b390"} Dec 06 07:17:43 crc kubenswrapper[4945]: I1206 07:17:43.879002 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6ccd85df7-25dfh" podStartSLOduration=2.878984739 podStartE2EDuration="2.878984739s" podCreationTimestamp="2025-12-06 07:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:43.875135484 +0000 UTC m=+1477.329996538" watchObservedRunningTime="2025-12-06 07:17:43.878984739 +0000 UTC m=+1477.333845773" Dec 06 07:17:43 crc kubenswrapper[4945]: I1206 07:17:43.917444 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.917419445 podStartE2EDuration="3.917419445s" podCreationTimestamp="2025-12-06 07:17:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:17:43.909595882 +0000 UTC m=+1477.364456926" watchObservedRunningTime="2025-12-06 07:17:43.917419445 +0000 UTC m=+1477.372280489" Dec 06 07:17:45 crc kubenswrapper[4945]: I1206 07:17:45.496671 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 07:17:45 crc kubenswrapper[4945]: I1206 07:17:45.875534 4945 generic.go:334] "Generic (PLEG): container finished" podID="72e9a7b2-b482-4130-819a-95ac1106e218" containerID="3ab75bae5874506e9ae523bd85a3d001e584020c211168290d995e46abd09b61" exitCode=0 Dec 06 07:17:45 crc kubenswrapper[4945]: I1206 07:17:45.876241 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72e9a7b2-b482-4130-819a-95ac1106e218","Type":"ContainerDied","Data":"3ab75bae5874506e9ae523bd85a3d001e584020c211168290d995e46abd09b61"} Dec 06 07:17:48 crc kubenswrapper[4945]: I1206 07:17:48.116267 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:17:48 crc kubenswrapper[4945]: I1206 07:17:48.116922 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5083dbae-2860-4efd-84e5-503b3a11d99b" containerName="glance-log" containerID="cri-o://c9f3cc29bda0b30b5a9c7145f3e4d1a71bf681a69aebb900b3ca0ec8bdb9d06b" gracePeriod=30 Dec 06 07:17:48 crc kubenswrapper[4945]: I1206 07:17:48.117019 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5083dbae-2860-4efd-84e5-503b3a11d99b" containerName="glance-httpd" containerID="cri-o://b513e7d8f6cdca78c4bbf0d6ee1dff4ffe512e4d9e1d9369b644275792ab4e24" gracePeriod=30 Dec 06 07:17:49 crc kubenswrapper[4945]: I1206 07:17:49.029412 4945 generic.go:334] "Generic (PLEG): container finished" podID="5083dbae-2860-4efd-84e5-503b3a11d99b" containerID="c9f3cc29bda0b30b5a9c7145f3e4d1a71bf681a69aebb900b3ca0ec8bdb9d06b" exitCode=143 Dec 06 07:17:49 crc kubenswrapper[4945]: I1206 07:17:49.029466 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5083dbae-2860-4efd-84e5-503b3a11d99b","Type":"ContainerDied","Data":"c9f3cc29bda0b30b5a9c7145f3e4d1a71bf681a69aebb900b3ca0ec8bdb9d06b"} Dec 06 07:17:50 crc kubenswrapper[4945]: I1206 07:17:50.853162 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:51.927181 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:51.927483 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8809238d-f467-4a48-b009-15c8b974aed0" containerName="glance-log" containerID="cri-o://b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5" gracePeriod=30 Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:51.927567 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8809238d-f467-4a48-b009-15c8b974aed0" containerName="glance-httpd" containerID="cri-o://46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79" gracePeriod=30 Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.063005 4945 generic.go:334] "Generic (PLEG): container finished" podID="5083dbae-2860-4efd-84e5-503b3a11d99b" containerID="b513e7d8f6cdca78c4bbf0d6ee1dff4ffe512e4d9e1d9369b644275792ab4e24" exitCode=0 Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.063073 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5083dbae-2860-4efd-84e5-503b3a11d99b","Type":"ContainerDied","Data":"b513e7d8f6cdca78c4bbf0d6ee1dff4ffe512e4d9e1d9369b644275792ab4e24"} Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.065242 4945 generic.go:334] "Generic (PLEG): container finished" podID="8809238d-f467-4a48-b009-15c8b974aed0" containerID="b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5" exitCode=143 Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.065274 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8809238d-f467-4a48-b009-15c8b974aed0","Type":"ContainerDied","Data":"b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5"} Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.195793 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.223990 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-log-httpd\") pod \"72e9a7b2-b482-4130-819a-95ac1106e218\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.224146 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-sg-core-conf-yaml\") pod \"72e9a7b2-b482-4130-819a-95ac1106e218\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.224214 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmqcw\" (UniqueName: \"kubernetes.io/projected/72e9a7b2-b482-4130-819a-95ac1106e218-kube-api-access-pmqcw\") pod \"72e9a7b2-b482-4130-819a-95ac1106e218\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.224239 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-scripts\") pod \"72e9a7b2-b482-4130-819a-95ac1106e218\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.224295 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-config-data\") pod \"72e9a7b2-b482-4130-819a-95ac1106e218\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.224340 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-combined-ca-bundle\") pod \"72e9a7b2-b482-4130-819a-95ac1106e218\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.224366 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-run-httpd\") pod \"72e9a7b2-b482-4130-819a-95ac1106e218\" (UID: \"72e9a7b2-b482-4130-819a-95ac1106e218\") " Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.227737 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "72e9a7b2-b482-4130-819a-95ac1106e218" (UID: "72e9a7b2-b482-4130-819a-95ac1106e218"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.229201 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "72e9a7b2-b482-4130-819a-95ac1106e218" (UID: "72e9a7b2-b482-4130-819a-95ac1106e218"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.248597 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-scripts" (OuterVolumeSpecName: "scripts") pod "72e9a7b2-b482-4130-819a-95ac1106e218" (UID: "72e9a7b2-b482-4130-819a-95ac1106e218"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.248624 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e9a7b2-b482-4130-819a-95ac1106e218-kube-api-access-pmqcw" (OuterVolumeSpecName: "kube-api-access-pmqcw") pod "72e9a7b2-b482-4130-819a-95ac1106e218" (UID: "72e9a7b2-b482-4130-819a-95ac1106e218"). InnerVolumeSpecName "kube-api-access-pmqcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.316811 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.323966 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.329294 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.329628 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmqcw\" (UniqueName: \"kubernetes.io/projected/72e9a7b2-b482-4130-819a-95ac1106e218-kube-api-access-pmqcw\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.329856 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.329946 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72e9a7b2-b482-4130-819a-95ac1106e218-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.337969 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "72e9a7b2-b482-4130-819a-95ac1106e218" (UID: "72e9a7b2-b482-4130-819a-95ac1106e218"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.386914 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72e9a7b2-b482-4130-819a-95ac1106e218" (UID: "72e9a7b2-b482-4130-819a-95ac1106e218"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.430752 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-config-data" (OuterVolumeSpecName: "config-data") pod "72e9a7b2-b482-4130-819a-95ac1106e218" (UID: "72e9a7b2-b482-4130-819a-95ac1106e218"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.433602 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.434902 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:52 crc kubenswrapper[4945]: I1206 07:17:52.434919 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72e9a7b2-b482-4130-819a-95ac1106e218-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.079861 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72e9a7b2-b482-4130-819a-95ac1106e218","Type":"ContainerDied","Data":"b067a2c52527c1fa782f2408d497d11103b1b4f6a63115a9e188fbf38a9ed35a"} Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.079918 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.079939 4945 scope.go:117] "RemoveContainer" containerID="479ab6973236d7632f68bb5b57d0888dc2c07de95ae8cfc1029ad9290416ff1d" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.130352 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.142985 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.160636 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:53 crc kubenswrapper[4945]: E1206 07:17:53.161176 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="ceilometer-notification-agent" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.161258 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="ceilometer-notification-agent" Dec 06 07:17:53 crc kubenswrapper[4945]: E1206 07:17:53.161307 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="sg-core" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.161317 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="sg-core" Dec 06 07:17:53 crc kubenswrapper[4945]: E1206 07:17:53.161346 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="proxy-httpd" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.161357 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="proxy-httpd" Dec 06 07:17:53 crc kubenswrapper[4945]: E1206 07:17:53.161380 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="ceilometer-central-agent" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.161387 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="ceilometer-central-agent" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.161608 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="proxy-httpd" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.161640 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="sg-core" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.161652 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="ceilometer-notification-agent" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.161662 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="ceilometer-central-agent" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.167764 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.174931 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.175008 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.179389 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.269043 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-run-httpd\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.269513 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2m8h\" (UniqueName: \"kubernetes.io/projected/a8cc641a-6a8f-4788-9194-4707a4a18951-kube-api-access-q2m8h\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.269574 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.270666 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.270737 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-log-httpd\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.270790 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-scripts\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.270821 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-config-data\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.373361 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.375188 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-log-httpd\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.375244 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-scripts\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.375267 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-config-data\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.375696 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-run-httpd\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.375747 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2m8h\" (UniqueName: \"kubernetes.io/projected/a8cc641a-6a8f-4788-9194-4707a4a18951-kube-api-access-q2m8h\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.375783 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.376198 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-log-httpd\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.376270 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-run-httpd\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.382252 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-config-data\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.383488 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-scripts\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.388897 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.390040 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.396814 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2m8h\" (UniqueName: \"kubernetes.io/projected/a8cc641a-6a8f-4788-9194-4707a4a18951-kube-api-access-q2m8h\") pod \"ceilometer-0\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.499448 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:17:53 crc kubenswrapper[4945]: I1206 07:17:53.677668 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:54 crc kubenswrapper[4945]: I1206 07:17:54.965015 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" path="/var/lib/kubelet/pods/72e9a7b2-b482-4130-819a-95ac1106e218/volumes" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.234246 4945 scope.go:117] "RemoveContainer" containerID="df489374460e08de26f48143dd596d770a90fe2c61e81c40c9410a4bcd4d5df6" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.258325 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.296586 4945 scope.go:117] "RemoveContainer" containerID="3ab75bae5874506e9ae523bd85a3d001e584020c211168290d995e46abd09b61" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.311747 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-logs\") pod \"5083dbae-2860-4efd-84e5-503b3a11d99b\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.311868 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"5083dbae-2860-4efd-84e5-503b3a11d99b\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.312083 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-httpd-run\") pod \"5083dbae-2860-4efd-84e5-503b3a11d99b\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.312144 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-config-data\") pod \"5083dbae-2860-4efd-84e5-503b3a11d99b\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.312180 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-combined-ca-bundle\") pod \"5083dbae-2860-4efd-84e5-503b3a11d99b\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.312251 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-scripts\") pod \"5083dbae-2860-4efd-84e5-503b3a11d99b\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.312437 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-public-tls-certs\") pod \"5083dbae-2860-4efd-84e5-503b3a11d99b\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.312482 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k89ct\" (UniqueName: \"kubernetes.io/projected/5083dbae-2860-4efd-84e5-503b3a11d99b-kube-api-access-k89ct\") pod \"5083dbae-2860-4efd-84e5-503b3a11d99b\" (UID: \"5083dbae-2860-4efd-84e5-503b3a11d99b\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.313168 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-logs" (OuterVolumeSpecName: "logs") pod "5083dbae-2860-4efd-84e5-503b3a11d99b" (UID: "5083dbae-2860-4efd-84e5-503b3a11d99b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.317437 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5083dbae-2860-4efd-84e5-503b3a11d99b" (UID: "5083dbae-2860-4efd-84e5-503b3a11d99b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.318446 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5083dbae-2860-4efd-84e5-503b3a11d99b-kube-api-access-k89ct" (OuterVolumeSpecName: "kube-api-access-k89ct") pod "5083dbae-2860-4efd-84e5-503b3a11d99b" (UID: "5083dbae-2860-4efd-84e5-503b3a11d99b"). InnerVolumeSpecName "kube-api-access-k89ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.332689 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "5083dbae-2860-4efd-84e5-503b3a11d99b" (UID: "5083dbae-2860-4efd-84e5-503b3a11d99b"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.345393 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-scripts" (OuterVolumeSpecName: "scripts") pod "5083dbae-2860-4efd-84e5-503b3a11d99b" (UID: "5083dbae-2860-4efd-84e5-503b3a11d99b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.393626 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5083dbae-2860-4efd-84e5-503b3a11d99b" (UID: "5083dbae-2860-4efd-84e5-503b3a11d99b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.415724 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.419359 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.419377 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.419389 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k89ct\" (UniqueName: \"kubernetes.io/projected/5083dbae-2860-4efd-84e5-503b3a11d99b-kube-api-access-k89ct\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.419403 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5083dbae-2860-4efd-84e5-503b3a11d99b-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.419428 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.471697 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.488444 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-config-data" (OuterVolumeSpecName: "config-data") pod "5083dbae-2860-4efd-84e5-503b3a11d99b" (UID: "5083dbae-2860-4efd-84e5-503b3a11d99b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.489751 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5083dbae-2860-4efd-84e5-503b3a11d99b" (UID: "5083dbae-2860-4efd-84e5-503b3a11d99b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.521359 4945 scope.go:117] "RemoveContainer" containerID="2139e668db9a92f5f7706eff1e6c504a57c97c94612cadb38beac36f2fa9ed10" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.524035 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.524061 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.524072 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5083dbae-2860-4efd-84e5-503b3a11d99b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.717649 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.830466 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-combined-ca-bundle\") pod \"8809238d-f467-4a48-b009-15c8b974aed0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.830598 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-httpd-run\") pod \"8809238d-f467-4a48-b009-15c8b974aed0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.830626 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-logs\") pod \"8809238d-f467-4a48-b009-15c8b974aed0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.830719 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-config-data\") pod \"8809238d-f467-4a48-b009-15c8b974aed0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.830780 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-scripts\") pod \"8809238d-f467-4a48-b009-15c8b974aed0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.830821 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-internal-tls-certs\") pod \"8809238d-f467-4a48-b009-15c8b974aed0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.830848 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"8809238d-f467-4a48-b009-15c8b974aed0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.830914 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5wfl\" (UniqueName: \"kubernetes.io/projected/8809238d-f467-4a48-b009-15c8b974aed0-kube-api-access-l5wfl\") pod \"8809238d-f467-4a48-b009-15c8b974aed0\" (UID: \"8809238d-f467-4a48-b009-15c8b974aed0\") " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.831479 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-logs" (OuterVolumeSpecName: "logs") pod "8809238d-f467-4a48-b009-15c8b974aed0" (UID: "8809238d-f467-4a48-b009-15c8b974aed0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.832097 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8809238d-f467-4a48-b009-15c8b974aed0" (UID: "8809238d-f467-4a48-b009-15c8b974aed0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.845010 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "8809238d-f467-4a48-b009-15c8b974aed0" (UID: "8809238d-f467-4a48-b009-15c8b974aed0"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.845143 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-scripts" (OuterVolumeSpecName: "scripts") pod "8809238d-f467-4a48-b009-15c8b974aed0" (UID: "8809238d-f467-4a48-b009-15c8b974aed0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.845265 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8809238d-f467-4a48-b009-15c8b974aed0-kube-api-access-l5wfl" (OuterVolumeSpecName: "kube-api-access-l5wfl") pod "8809238d-f467-4a48-b009-15c8b974aed0" (UID: "8809238d-f467-4a48-b009-15c8b974aed0"). InnerVolumeSpecName "kube-api-access-l5wfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.870853 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8809238d-f467-4a48-b009-15c8b974aed0" (UID: "8809238d-f467-4a48-b009-15c8b974aed0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.894728 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-config-data" (OuterVolumeSpecName: "config-data") pod "8809238d-f467-4a48-b009-15c8b974aed0" (UID: "8809238d-f467-4a48-b009-15c8b974aed0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.910232 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8809238d-f467-4a48-b009-15c8b974aed0" (UID: "8809238d-f467-4a48-b009-15c8b974aed0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.912871 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.934719 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5wfl\" (UniqueName: \"kubernetes.io/projected/8809238d-f467-4a48-b009-15c8b974aed0-kube-api-access-l5wfl\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.934782 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.934805 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.934824 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8809238d-f467-4a48-b009-15c8b974aed0-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.934835 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.934846 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.934856 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8809238d-f467-4a48-b009-15c8b974aed0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.934911 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 06 07:17:55 crc kubenswrapper[4945]: I1206 07:17:55.956147 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.036915 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.108998 4945 generic.go:334] "Generic (PLEG): container finished" podID="8809238d-f467-4a48-b009-15c8b974aed0" containerID="46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79" exitCode=0 Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.109113 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.109117 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8809238d-f467-4a48-b009-15c8b974aed0","Type":"ContainerDied","Data":"46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79"} Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.109199 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8809238d-f467-4a48-b009-15c8b974aed0","Type":"ContainerDied","Data":"d49d97c9a969ed79389bad7a9f3619af73c623f420d9fcf018d0173695c908fc"} Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.109226 4945 scope.go:117] "RemoveContainer" containerID="46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.114019 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"7e90dc2b-f137-411b-87e4-b46f36d3034e","Type":"ContainerStarted","Data":"ff0585a6020bd13a78632832e2192d413450d1f996c3ce1657896dcee244dbdb"} Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.117501 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8cc641a-6a8f-4788-9194-4707a4a18951","Type":"ContainerStarted","Data":"edb7c49e32dc51eabba79932fdfac6bd39aa389ec8f83d17f57707416fb15cec"} Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.130245 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5083dbae-2860-4efd-84e5-503b3a11d99b","Type":"ContainerDied","Data":"19599e6d8c229bc6c5cbe70fd54ceb47dd313d99909ad7f2c8e98b518d9fe21d"} Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.130375 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.144864 4945 scope.go:117] "RemoveContainer" containerID="b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.147841 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.459933929 podStartE2EDuration="20.147817382s" podCreationTimestamp="2025-12-06 07:17:36 +0000 UTC" firstStartedPulling="2025-12-06 07:17:37.609045469 +0000 UTC m=+1471.063906513" lastFinishedPulling="2025-12-06 07:17:55.296928922 +0000 UTC m=+1488.751789966" observedRunningTime="2025-12-06 07:17:56.134846041 +0000 UTC m=+1489.589707095" watchObservedRunningTime="2025-12-06 07:17:56.147817382 +0000 UTC m=+1489.602678426" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.178484 4945 scope.go:117] "RemoveContainer" containerID="46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79" Dec 06 07:17:56 crc kubenswrapper[4945]: E1206 07:17:56.182559 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79\": container with ID starting with 46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79 not found: ID does not exist" containerID="46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.182603 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79"} err="failed to get container status \"46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79\": rpc error: code = NotFound desc = could not find container \"46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79\": container with ID starting with 46e19371f6252c85d39c22c1d94ae98a2691c85e46b7bbdf745fe7528237de79 not found: ID does not exist" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.182632 4945 scope.go:117] "RemoveContainer" containerID="b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5" Dec 06 07:17:56 crc kubenswrapper[4945]: E1206 07:17:56.183743 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5\": container with ID starting with b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5 not found: ID does not exist" containerID="b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.183766 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5"} err="failed to get container status \"b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5\": rpc error: code = NotFound desc = could not find container \"b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5\": container with ID starting with b139e6277f7e3abd7b271c583f41b679fe54d5bfefc07c1ba0ab690e7cf6e5e5 not found: ID does not exist" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.183784 4945 scope.go:117] "RemoveContainer" containerID="b513e7d8f6cdca78c4bbf0d6ee1dff4ffe512e4d9e1d9369b644275792ab4e24" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.184773 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.201394 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.210436 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.217360 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.244111 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:17:56 crc kubenswrapper[4945]: E1206 07:17:56.244570 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8809238d-f467-4a48-b009-15c8b974aed0" containerName="glance-httpd" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.244587 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8809238d-f467-4a48-b009-15c8b974aed0" containerName="glance-httpd" Dec 06 07:17:56 crc kubenswrapper[4945]: E1206 07:17:56.244598 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5083dbae-2860-4efd-84e5-503b3a11d99b" containerName="glance-log" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.244605 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5083dbae-2860-4efd-84e5-503b3a11d99b" containerName="glance-log" Dec 06 07:17:56 crc kubenswrapper[4945]: E1206 07:17:56.244627 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5083dbae-2860-4efd-84e5-503b3a11d99b" containerName="glance-httpd" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.244634 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5083dbae-2860-4efd-84e5-503b3a11d99b" containerName="glance-httpd" Dec 06 07:17:56 crc kubenswrapper[4945]: E1206 07:17:56.244647 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8809238d-f467-4a48-b009-15c8b974aed0" containerName="glance-log" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.244655 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8809238d-f467-4a48-b009-15c8b974aed0" containerName="glance-log" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.244853 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5083dbae-2860-4efd-84e5-503b3a11d99b" containerName="glance-log" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.244871 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5083dbae-2860-4efd-84e5-503b3a11d99b" containerName="glance-httpd" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.244882 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8809238d-f467-4a48-b009-15c8b974aed0" containerName="glance-log" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.244903 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8809238d-f467-4a48-b009-15c8b974aed0" containerName="glance-httpd" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.245867 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.274457 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.275685 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.276156 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5m642" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.291268 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.322159 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.324721 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.328594 4945 scope.go:117] "RemoveContainer" containerID="c9f3cc29bda0b30b5a9c7145f3e4d1a71bf681a69aebb900b3ca0ec8bdb9d06b" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.331531 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.331861 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.355353 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.365927 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.449001 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.449062 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.449181 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.449209 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.449230 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.449264 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-logs\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.449313 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljgs5\" (UniqueName: \"kubernetes.io/projected/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-kube-api-access-ljgs5\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.449349 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.551845 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.551965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.551997 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552021 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552040 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552063 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552088 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552150 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552227 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552259 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552310 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-logs\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552343 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552385 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-logs\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552421 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552451 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljgs5\" (UniqueName: \"kubernetes.io/projected/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-kube-api-access-ljgs5\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.552495 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s75pv\" (UniqueName: \"kubernetes.io/projected/8ba3e057-d321-48ca-b7a0-8d791cbb7169-kube-api-access-s75pv\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.553728 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.554013 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-logs\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.554234 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.558913 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.568060 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.568128 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.571979 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.577214 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljgs5\" (UniqueName: \"kubernetes.io/projected/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-kube-api-access-ljgs5\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.601733 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.654397 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-logs\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.654490 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.654542 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s75pv\" (UniqueName: \"kubernetes.io/projected/8ba3e057-d321-48ca-b7a0-8d791cbb7169-kube-api-access-s75pv\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.654623 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.654651 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.654683 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.654713 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.654763 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.655153 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.655241 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.655555 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-logs\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.661311 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.661605 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.661924 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.662076 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.679804 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s75pv\" (UniqueName: \"kubernetes.io/projected/8ba3e057-d321-48ca-b7a0-8d791cbb7169-kube-api-access-s75pv\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.695637 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.898082 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.946848 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.964926 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5083dbae-2860-4efd-84e5-503b3a11d99b" path="/var/lib/kubelet/pods/5083dbae-2860-4efd-84e5-503b3a11d99b/volumes" Dec 06 07:17:56 crc kubenswrapper[4945]: I1206 07:17:56.965754 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8809238d-f467-4a48-b009-15c8b974aed0" path="/var/lib/kubelet/pods/8809238d-f467-4a48-b009-15c8b974aed0/volumes" Dec 06 07:17:57 crc kubenswrapper[4945]: I1206 07:17:57.172343 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8cc641a-6a8f-4788-9194-4707a4a18951","Type":"ContainerStarted","Data":"ed34589057817ad72a5317eab6b9f21905263e7d0873427981949caa82987191"} Dec 06 07:17:57 crc kubenswrapper[4945]: I1206 07:17:57.589545 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:17:57 crc kubenswrapper[4945]: I1206 07:17:57.761835 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.240829 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8cc641a-6a8f-4788-9194-4707a4a18951","Type":"ContainerStarted","Data":"df6d926343b2c56b8535b92675f93512dbdb6c74f7bdf758728c976bec5fd4ce"} Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.241369 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8cc641a-6a8f-4788-9194-4707a4a18951","Type":"ContainerStarted","Data":"79de2a8c48d8cdf9a756bedc61a029ecea66e0808f3605f405bca9e9052a6abf"} Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.249541 4945 generic.go:334] "Generic (PLEG): container finished" podID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" containerID="b5d25d4b24441a520223e8b318d2e0dca92158e2459d09ca919b3aafc1b0d487" exitCode=137 Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.249608 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a1acab70-a94a-4bfb-878e-a20c63b6c7f2","Type":"ContainerDied","Data":"b5d25d4b24441a520223e8b318d2e0dca92158e2459d09ca919b3aafc1b0d487"} Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.253460 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ba3e057-d321-48ca-b7a0-8d791cbb7169","Type":"ContainerStarted","Data":"e3437617f02eb68845c8942f8f14070ba5170613d8bd7ab6f091ee633fae3a22"} Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.265610 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf0fa4a-4a9c-4644-9fe3-413500c59a98","Type":"ContainerStarted","Data":"1f62cdaba764c253ae0c59729461040b63726374f0f788c093cebdb6d800f44a"} Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.441650 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.604622 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-scripts\") pod \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.604914 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data-custom\") pod \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.604945 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9qqg\" (UniqueName: \"kubernetes.io/projected/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-kube-api-access-r9qqg\") pod \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.605081 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-logs\") pod \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.605126 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data\") pod \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.605182 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-combined-ca-bundle\") pod \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.605219 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-etc-machine-id\") pod \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\" (UID: \"a1acab70-a94a-4bfb-878e-a20c63b6c7f2\") " Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.605701 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a1acab70-a94a-4bfb-878e-a20c63b6c7f2" (UID: "a1acab70-a94a-4bfb-878e-a20c63b6c7f2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.607253 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-logs" (OuterVolumeSpecName: "logs") pod "a1acab70-a94a-4bfb-878e-a20c63b6c7f2" (UID: "a1acab70-a94a-4bfb-878e-a20c63b6c7f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.614424 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a1acab70-a94a-4bfb-878e-a20c63b6c7f2" (UID: "a1acab70-a94a-4bfb-878e-a20c63b6c7f2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.614672 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-scripts" (OuterVolumeSpecName: "scripts") pod "a1acab70-a94a-4bfb-878e-a20c63b6c7f2" (UID: "a1acab70-a94a-4bfb-878e-a20c63b6c7f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.630638 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-kube-api-access-r9qqg" (OuterVolumeSpecName: "kube-api-access-r9qqg") pod "a1acab70-a94a-4bfb-878e-a20c63b6c7f2" (UID: "a1acab70-a94a-4bfb-878e-a20c63b6c7f2"). InnerVolumeSpecName "kube-api-access-r9qqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.645324 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1acab70-a94a-4bfb-878e-a20c63b6c7f2" (UID: "a1acab70-a94a-4bfb-878e-a20c63b6c7f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.679895 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data" (OuterVolumeSpecName: "config-data") pod "a1acab70-a94a-4bfb-878e-a20c63b6c7f2" (UID: "a1acab70-a94a-4bfb-878e-a20c63b6c7f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.707525 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.707564 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.707578 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9qqg\" (UniqueName: \"kubernetes.io/projected/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-kube-api-access-r9qqg\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.707593 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.707603 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.707613 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:58 crc kubenswrapper[4945]: I1206 07:17:58.707623 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1acab70-a94a-4bfb-878e-a20c63b6c7f2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.286161 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ba3e057-d321-48ca-b7a0-8d791cbb7169","Type":"ContainerStarted","Data":"41457071a0e5ace0cbbf96b7adf5ba3e23ee9463d1013d0f9d4b4f9916e1e917"} Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.288974 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf0fa4a-4a9c-4644-9fe3-413500c59a98","Type":"ContainerStarted","Data":"31c19a7223a90ce2817b04d80780886488d735a425cbc88ccf01eeac5899bab6"} Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.293111 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a1acab70-a94a-4bfb-878e-a20c63b6c7f2","Type":"ContainerDied","Data":"caf7c3b46300aa2dd0e198223a99a610ef7571d05e539054a5d7f5d2af41eb1b"} Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.293166 4945 scope.go:117] "RemoveContainer" containerID="b5d25d4b24441a520223e8b318d2e0dca92158e2459d09ca919b3aafc1b0d487" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.293523 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.331388 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.369523 4945 scope.go:117] "RemoveContainer" containerID="a2852e09c45fa23e140608ec02c569334b53cf80a99edae75b75522c7d48f454" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.373701 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.415170 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:17:59 crc kubenswrapper[4945]: E1206 07:17:59.415768 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" containerName="cinder-api-log" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.415793 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" containerName="cinder-api-log" Dec 06 07:17:59 crc kubenswrapper[4945]: E1206 07:17:59.415822 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" containerName="cinder-api" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.415831 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" containerName="cinder-api" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.416035 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" containerName="cinder-api" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.416061 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" containerName="cinder-api-log" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.417123 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.425210 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.425408 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.425440 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.429130 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.525460 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data-custom\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.525759 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.525831 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edbffd82-c356-43a3-94a3-3549b7d9cd31-logs\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.525951 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.526008 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-public-tls-certs\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.526029 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-scripts\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.526069 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.526088 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m595p\" (UniqueName: \"kubernetes.io/projected/edbffd82-c356-43a3-94a3-3549b7d9cd31-kube-api-access-m595p\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.526108 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edbffd82-c356-43a3-94a3-3549b7d9cd31-etc-machine-id\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.628149 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.628221 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-public-tls-certs\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.628240 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-scripts\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.628271 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.628307 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m595p\" (UniqueName: \"kubernetes.io/projected/edbffd82-c356-43a3-94a3-3549b7d9cd31-kube-api-access-m595p\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.628326 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edbffd82-c356-43a3-94a3-3549b7d9cd31-etc-machine-id\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.628388 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data-custom\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.628412 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.628468 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edbffd82-c356-43a3-94a3-3549b7d9cd31-logs\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.628865 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edbffd82-c356-43a3-94a3-3549b7d9cd31-logs\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.629049 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edbffd82-c356-43a3-94a3-3549b7d9cd31-etc-machine-id\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.635060 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.635910 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-public-tls-certs\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.637097 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data-custom\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.637541 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.644421 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.644936 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-scripts\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.648636 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m595p\" (UniqueName: \"kubernetes.io/projected/edbffd82-c356-43a3-94a3-3549b7d9cd31-kube-api-access-m595p\") pod \"cinder-api-0\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " pod="openstack/cinder-api-0" Dec 06 07:17:59 crc kubenswrapper[4945]: I1206 07:17:59.753710 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.066625 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.323568 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf0fa4a-4a9c-4644-9fe3-413500c59a98","Type":"ContainerStarted","Data":"f1978a8cc52c8c7435676004e144a0b8373ccf11fd0f8cada610a31d96b4bf20"} Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.326763 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"edbffd82-c356-43a3-94a3-3549b7d9cd31","Type":"ContainerStarted","Data":"7649c561156cfdd3269bbf7b7920020626421572c2eafcd315a0c2aebc66ac69"} Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.331665 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8cc641a-6a8f-4788-9194-4707a4a18951","Type":"ContainerStarted","Data":"f7492d5e5eac181374232bb84375c7fd72e56b1450784144d7ba040bdd005f45"} Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.331739 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="ceilometer-central-agent" containerID="cri-o://ed34589057817ad72a5317eab6b9f21905263e7d0873427981949caa82987191" gracePeriod=30 Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.331763 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="sg-core" containerID="cri-o://df6d926343b2c56b8535b92675f93512dbdb6c74f7bdf758728c976bec5fd4ce" gracePeriod=30 Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.331786 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.331784 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="ceilometer-notification-agent" containerID="cri-o://79de2a8c48d8cdf9a756bedc61a029ecea66e0808f3605f405bca9e9052a6abf" gracePeriod=30 Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.331844 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="proxy-httpd" containerID="cri-o://f7492d5e5eac181374232bb84375c7fd72e56b1450784144d7ba040bdd005f45" gracePeriod=30 Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.347180 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ba3e057-d321-48ca-b7a0-8d791cbb7169","Type":"ContainerStarted","Data":"f75d79cb35ec6aeed2fbb5df193689624fa68ec6f36d6f07600634a6a1eb3d4f"} Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.367813 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.367793949 podStartE2EDuration="4.367793949s" podCreationTimestamp="2025-12-06 07:17:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:18:00.356885004 +0000 UTC m=+1493.811746048" watchObservedRunningTime="2025-12-06 07:18:00.367793949 +0000 UTC m=+1493.822654993" Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.391981 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.791437085 podStartE2EDuration="7.391957423s" podCreationTimestamp="2025-12-06 07:17:53 +0000 UTC" firstStartedPulling="2025-12-06 07:17:55.921152611 +0000 UTC m=+1489.376013655" lastFinishedPulling="2025-12-06 07:17:59.521672949 +0000 UTC m=+1492.976533993" observedRunningTime="2025-12-06 07:18:00.378243502 +0000 UTC m=+1493.833104546" watchObservedRunningTime="2025-12-06 07:18:00.391957423 +0000 UTC m=+1493.846818467" Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.434027 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.433999051 podStartE2EDuration="4.433999051s" podCreationTimestamp="2025-12-06 07:17:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:18:00.40957813 +0000 UTC m=+1493.864439184" watchObservedRunningTime="2025-12-06 07:18:00.433999051 +0000 UTC m=+1493.888860105" Dec 06 07:18:00 crc kubenswrapper[4945]: I1206 07:18:00.976086 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1acab70-a94a-4bfb-878e-a20c63b6c7f2" path="/var/lib/kubelet/pods/a1acab70-a94a-4bfb-878e-a20c63b6c7f2/volumes" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.370527 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"edbffd82-c356-43a3-94a3-3549b7d9cd31","Type":"ContainerStarted","Data":"c3b041158b4f4c593ae33438ee22c67b72a78b4734276f3449241b8f0221a4dd"} Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.390629 4945 generic.go:334] "Generic (PLEG): container finished" podID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerID="f7492d5e5eac181374232bb84375c7fd72e56b1450784144d7ba040bdd005f45" exitCode=0 Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.390662 4945 generic.go:334] "Generic (PLEG): container finished" podID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerID="df6d926343b2c56b8535b92675f93512dbdb6c74f7bdf758728c976bec5fd4ce" exitCode=2 Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.390672 4945 generic.go:334] "Generic (PLEG): container finished" podID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerID="79de2a8c48d8cdf9a756bedc61a029ecea66e0808f3605f405bca9e9052a6abf" exitCode=0 Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.390680 4945 generic.go:334] "Generic (PLEG): container finished" podID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerID="ed34589057817ad72a5317eab6b9f21905263e7d0873427981949caa82987191" exitCode=0 Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.390732 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8cc641a-6a8f-4788-9194-4707a4a18951","Type":"ContainerDied","Data":"f7492d5e5eac181374232bb84375c7fd72e56b1450784144d7ba040bdd005f45"} Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.390806 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8cc641a-6a8f-4788-9194-4707a4a18951","Type":"ContainerDied","Data":"df6d926343b2c56b8535b92675f93512dbdb6c74f7bdf758728c976bec5fd4ce"} Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.390817 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8cc641a-6a8f-4788-9194-4707a4a18951","Type":"ContainerDied","Data":"79de2a8c48d8cdf9a756bedc61a029ecea66e0808f3605f405bca9e9052a6abf"} Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.390827 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8cc641a-6a8f-4788-9194-4707a4a18951","Type":"ContainerDied","Data":"ed34589057817ad72a5317eab6b9f21905263e7d0873427981949caa82987191"} Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.390839 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8cc641a-6a8f-4788-9194-4707a4a18951","Type":"ContainerDied","Data":"edb7c49e32dc51eabba79932fdfac6bd39aa389ec8f83d17f57707416fb15cec"} Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.390850 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edb7c49e32dc51eabba79932fdfac6bd39aa389ec8f83d17f57707416fb15cec" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.397421 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.566876 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-scripts\") pod \"a8cc641a-6a8f-4788-9194-4707a4a18951\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.566933 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-run-httpd\") pod \"a8cc641a-6a8f-4788-9194-4707a4a18951\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.566968 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-combined-ca-bundle\") pod \"a8cc641a-6a8f-4788-9194-4707a4a18951\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.567036 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-sg-core-conf-yaml\") pod \"a8cc641a-6a8f-4788-9194-4707a4a18951\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.567074 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-log-httpd\") pod \"a8cc641a-6a8f-4788-9194-4707a4a18951\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.567341 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-config-data\") pod \"a8cc641a-6a8f-4788-9194-4707a4a18951\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.567376 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2m8h\" (UniqueName: \"kubernetes.io/projected/a8cc641a-6a8f-4788-9194-4707a4a18951-kube-api-access-q2m8h\") pod \"a8cc641a-6a8f-4788-9194-4707a4a18951\" (UID: \"a8cc641a-6a8f-4788-9194-4707a4a18951\") " Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.568904 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a8cc641a-6a8f-4788-9194-4707a4a18951" (UID: "a8cc641a-6a8f-4788-9194-4707a4a18951"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.569236 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a8cc641a-6a8f-4788-9194-4707a4a18951" (UID: "a8cc641a-6a8f-4788-9194-4707a4a18951"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.574642 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8cc641a-6a8f-4788-9194-4707a4a18951-kube-api-access-q2m8h" (OuterVolumeSpecName: "kube-api-access-q2m8h") pod "a8cc641a-6a8f-4788-9194-4707a4a18951" (UID: "a8cc641a-6a8f-4788-9194-4707a4a18951"). InnerVolumeSpecName "kube-api-access-q2m8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.583598 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-scripts" (OuterVolumeSpecName: "scripts") pod "a8cc641a-6a8f-4788-9194-4707a4a18951" (UID: "a8cc641a-6a8f-4788-9194-4707a4a18951"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.617479 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a8cc641a-6a8f-4788-9194-4707a4a18951" (UID: "a8cc641a-6a8f-4788-9194-4707a4a18951"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.656481 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8cc641a-6a8f-4788-9194-4707a4a18951" (UID: "a8cc641a-6a8f-4788-9194-4707a4a18951"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.670128 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.670163 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.670174 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.670186 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.670195 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8cc641a-6a8f-4788-9194-4707a4a18951-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.670210 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2m8h\" (UniqueName: \"kubernetes.io/projected/a8cc641a-6a8f-4788-9194-4707a4a18951-kube-api-access-q2m8h\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.701413 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-config-data" (OuterVolumeSpecName: "config-data") pod "a8cc641a-6a8f-4788-9194-4707a4a18951" (UID: "a8cc641a-6a8f-4788-9194-4707a4a18951"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:01 crc kubenswrapper[4945]: I1206 07:18:01.772618 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8cc641a-6a8f-4788-9194-4707a4a18951-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.402620 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"edbffd82-c356-43a3-94a3-3549b7d9cd31","Type":"ContainerStarted","Data":"b957bb81cfc39bf3a8747b1e1804ba1ec5dcb33d0889082d694cf4ffd810ef17"} Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.403886 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.403970 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.427062 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.427036361 podStartE2EDuration="3.427036361s" podCreationTimestamp="2025-12-06 07:17:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:18:02.423041742 +0000 UTC m=+1495.877902786" watchObservedRunningTime="2025-12-06 07:18:02.427036361 +0000 UTC m=+1495.881897405" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.455935 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.469751 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.498592 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:02 crc kubenswrapper[4945]: E1206 07:18:02.499089 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="ceilometer-central-agent" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.499108 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="ceilometer-central-agent" Dec 06 07:18:02 crc kubenswrapper[4945]: E1206 07:18:02.499123 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="ceilometer-notification-agent" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.499131 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="ceilometer-notification-agent" Dec 06 07:18:02 crc kubenswrapper[4945]: E1206 07:18:02.499150 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="proxy-httpd" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.499157 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="proxy-httpd" Dec 06 07:18:02 crc kubenswrapper[4945]: E1206 07:18:02.499169 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="sg-core" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.499175 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="sg-core" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.499372 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="ceilometer-notification-agent" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.499389 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="sg-core" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.499400 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="ceilometer-central-agent" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.499417 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" containerName="proxy-httpd" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.501225 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.504767 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.516353 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.522493 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.600066 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.600124 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.600147 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-log-httpd\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.600190 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-run-httpd\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.600323 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcsj2\" (UniqueName: \"kubernetes.io/projected/db588a5a-1e6e-436e-90ab-43c6a682fc2b-kube-api-access-lcsj2\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.600360 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-scripts\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.600384 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-config-data\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.624331 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:02 crc kubenswrapper[4945]: E1206 07:18:02.625590 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-lcsj2 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="db588a5a-1e6e-436e-90ab-43c6a682fc2b" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.702376 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.702434 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.702454 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-log-httpd\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.702488 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-run-httpd\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.702521 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcsj2\" (UniqueName: \"kubernetes.io/projected/db588a5a-1e6e-436e-90ab-43c6a682fc2b-kube-api-access-lcsj2\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.702549 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-scripts\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.702570 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-config-data\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.703502 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-run-httpd\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.704152 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-log-httpd\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.707002 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.707435 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-config-data\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.707794 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.714022 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-scripts\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.720034 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcsj2\" (UniqueName: \"kubernetes.io/projected/db588a5a-1e6e-436e-90ab-43c6a682fc2b-kube-api-access-lcsj2\") pod \"ceilometer-0\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " pod="openstack/ceilometer-0" Dec 06 07:18:02 crc kubenswrapper[4945]: I1206 07:18:02.964744 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8cc641a-6a8f-4788-9194-4707a4a18951" path="/var/lib/kubelet/pods/a8cc641a-6a8f-4788-9194-4707a4a18951/volumes" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.413203 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.429688 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.520062 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-config-data\") pod \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.520250 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcsj2\" (UniqueName: \"kubernetes.io/projected/db588a5a-1e6e-436e-90ab-43c6a682fc2b-kube-api-access-lcsj2\") pod \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.520308 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-scripts\") pod \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.520735 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "db588a5a-1e6e-436e-90ab-43c6a682fc2b" (UID: "db588a5a-1e6e-436e-90ab-43c6a682fc2b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.520792 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-log-httpd\") pod \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.520909 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-combined-ca-bundle\") pod \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.521356 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-run-httpd\") pod \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.521415 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-sg-core-conf-yaml\") pod \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\" (UID: \"db588a5a-1e6e-436e-90ab-43c6a682fc2b\") " Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.522437 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "db588a5a-1e6e-436e-90ab-43c6a682fc2b" (UID: "db588a5a-1e6e-436e-90ab-43c6a682fc2b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.525330 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.525362 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db588a5a-1e6e-436e-90ab-43c6a682fc2b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.525649 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-scripts" (OuterVolumeSpecName: "scripts") pod "db588a5a-1e6e-436e-90ab-43c6a682fc2b" (UID: "db588a5a-1e6e-436e-90ab-43c6a682fc2b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.526448 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-config-data" (OuterVolumeSpecName: "config-data") pod "db588a5a-1e6e-436e-90ab-43c6a682fc2b" (UID: "db588a5a-1e6e-436e-90ab-43c6a682fc2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.527552 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "db588a5a-1e6e-436e-90ab-43c6a682fc2b" (UID: "db588a5a-1e6e-436e-90ab-43c6a682fc2b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.528486 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db588a5a-1e6e-436e-90ab-43c6a682fc2b" (UID: "db588a5a-1e6e-436e-90ab-43c6a682fc2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.538829 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db588a5a-1e6e-436e-90ab-43c6a682fc2b-kube-api-access-lcsj2" (OuterVolumeSpecName: "kube-api-access-lcsj2") pod "db588a5a-1e6e-436e-90ab-43c6a682fc2b" (UID: "db588a5a-1e6e-436e-90ab-43c6a682fc2b"). InnerVolumeSpecName "kube-api-access-lcsj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.627127 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcsj2\" (UniqueName: \"kubernetes.io/projected/db588a5a-1e6e-436e-90ab-43c6a682fc2b-kube-api-access-lcsj2\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.627157 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.627167 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.627175 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:03 crc kubenswrapper[4945]: I1206 07:18:03.627183 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db588a5a-1e6e-436e-90ab-43c6a682fc2b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.426056 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.494326 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.513562 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.527218 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.530053 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.532095 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.533072 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.547375 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.648776 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-run-httpd\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.648823 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.648856 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.648885 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-config-data\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.648909 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-log-httpd\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.649025 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-scripts\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.649065 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-425g7\" (UniqueName: \"kubernetes.io/projected/57333dda-f06f-4d53-86ee-8514ff7151ea-kube-api-access-425g7\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.751374 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-scripts\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.751427 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-425g7\" (UniqueName: \"kubernetes.io/projected/57333dda-f06f-4d53-86ee-8514ff7151ea-kube-api-access-425g7\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.751518 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-run-httpd\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.751554 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.751593 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.751638 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-config-data\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.751680 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-log-httpd\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.752170 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-run-httpd\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.752657 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-log-httpd\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.757542 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.757694 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.757751 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-scripts\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.758116 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-config-data\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.770452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-425g7\" (UniqueName: \"kubernetes.io/projected/57333dda-f06f-4d53-86ee-8514ff7151ea-kube-api-access-425g7\") pod \"ceilometer-0\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.854506 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:04 crc kubenswrapper[4945]: I1206 07:18:04.973028 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db588a5a-1e6e-436e-90ab-43c6a682fc2b" path="/var/lib/kubelet/pods/db588a5a-1e6e-436e-90ab-43c6a682fc2b/volumes" Dec 06 07:18:05 crc kubenswrapper[4945]: I1206 07:18:05.301904 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:05 crc kubenswrapper[4945]: I1206 07:18:05.438112 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57333dda-f06f-4d53-86ee-8514ff7151ea","Type":"ContainerStarted","Data":"9bb626a8ff6b93dc38fef4868ad301c7fb337a8953d150c0d30d0f318506671d"} Dec 06 07:18:06 crc kubenswrapper[4945]: I1206 07:18:06.450539 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57333dda-f06f-4d53-86ee-8514ff7151ea","Type":"ContainerStarted","Data":"024ef6e2c14bea89f70ebb385437b4ea6f8705e27f84943de8f3ea92ab4132e1"} Dec 06 07:18:06 crc kubenswrapper[4945]: I1206 07:18:06.898302 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 07:18:06 crc kubenswrapper[4945]: I1206 07:18:06.898690 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 07:18:06 crc kubenswrapper[4945]: I1206 07:18:06.937978 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 07:18:06 crc kubenswrapper[4945]: I1206 07:18:06.947552 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 07:18:06 crc kubenswrapper[4945]: I1206 07:18:06.947601 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 07:18:07 crc kubenswrapper[4945]: I1206 07:18:07.008332 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 07:18:07 crc kubenswrapper[4945]: I1206 07:18:07.008407 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 07:18:07 crc kubenswrapper[4945]: I1206 07:18:07.021609 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 07:18:07 crc kubenswrapper[4945]: I1206 07:18:07.469484 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57333dda-f06f-4d53-86ee-8514ff7151ea","Type":"ContainerStarted","Data":"bbfb74902890e9a61b24d3f56f0e7a8bb54514b8af03f6d9cb65ceac20b20159"} Dec 06 07:18:07 crc kubenswrapper[4945]: I1206 07:18:07.469549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57333dda-f06f-4d53-86ee-8514ff7151ea","Type":"ContainerStarted","Data":"cfacc27b96dd10b0742f177f8e6346a834f8709ab3a4aea54db05a0c7502f042"} Dec 06 07:18:07 crc kubenswrapper[4945]: I1206 07:18:07.470476 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 07:18:07 crc kubenswrapper[4945]: I1206 07:18:07.470520 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 07:18:07 crc kubenswrapper[4945]: I1206 07:18:07.470532 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 07:18:07 crc kubenswrapper[4945]: I1206 07:18:07.473961 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.051324 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-lstm6"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.052927 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lstm6" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.089231 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lstm6"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.123127 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-operator-scripts\") pod \"nova-api-db-create-lstm6\" (UID: \"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd\") " pod="openstack/nova-api-db-create-lstm6" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.123186 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thkf6\" (UniqueName: \"kubernetes.io/projected/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-kube-api-access-thkf6\") pod \"nova-api-db-create-lstm6\" (UID: \"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd\") " pod="openstack/nova-api-db-create-lstm6" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.149671 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-zxg5j"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.151262 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zxg5j" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.177997 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-zxg5j"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.225847 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6stsb\" (UniqueName: \"kubernetes.io/projected/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-kube-api-access-6stsb\") pod \"nova-cell0-db-create-zxg5j\" (UID: \"7ce3702d-5de0-469b-a8e8-b531c4ea5afa\") " pod="openstack/nova-cell0-db-create-zxg5j" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.225897 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-operator-scripts\") pod \"nova-cell0-db-create-zxg5j\" (UID: \"7ce3702d-5de0-469b-a8e8-b531c4ea5afa\") " pod="openstack/nova-cell0-db-create-zxg5j" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.225941 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-operator-scripts\") pod \"nova-api-db-create-lstm6\" (UID: \"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd\") " pod="openstack/nova-api-db-create-lstm6" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.225970 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thkf6\" (UniqueName: \"kubernetes.io/projected/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-kube-api-access-thkf6\") pod \"nova-api-db-create-lstm6\" (UID: \"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd\") " pod="openstack/nova-api-db-create-lstm6" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.232928 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-operator-scripts\") pod \"nova-api-db-create-lstm6\" (UID: \"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd\") " pod="openstack/nova-api-db-create-lstm6" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.264024 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-badc-account-create-update-njdjh"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.265337 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-badc-account-create-update-njdjh" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.265474 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thkf6\" (UniqueName: \"kubernetes.io/projected/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-kube-api-access-thkf6\") pod \"nova-api-db-create-lstm6\" (UID: \"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd\") " pod="openstack/nova-api-db-create-lstm6" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.275641 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.283025 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-f5zbb"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.286189 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-f5zbb" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.307566 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-badc-account-create-update-njdjh"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.327244 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-operator-scripts\") pod \"nova-cell0-db-create-zxg5j\" (UID: \"7ce3702d-5de0-469b-a8e8-b531c4ea5afa\") " pod="openstack/nova-cell0-db-create-zxg5j" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.327335 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95a49175-2149-4cd4-88fc-e25ca5466ff0-operator-scripts\") pod \"nova-api-badc-account-create-update-njdjh\" (UID: \"95a49175-2149-4cd4-88fc-e25ca5466ff0\") " pod="openstack/nova-api-badc-account-create-update-njdjh" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.327400 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-operator-scripts\") pod \"nova-cell1-db-create-f5zbb\" (UID: \"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553\") " pod="openstack/nova-cell1-db-create-f5zbb" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.327441 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh5kl\" (UniqueName: \"kubernetes.io/projected/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-kube-api-access-sh5kl\") pod \"nova-cell1-db-create-f5zbb\" (UID: \"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553\") " pod="openstack/nova-cell1-db-create-f5zbb" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.327465 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb8k2\" (UniqueName: \"kubernetes.io/projected/95a49175-2149-4cd4-88fc-e25ca5466ff0-kube-api-access-fb8k2\") pod \"nova-api-badc-account-create-update-njdjh\" (UID: \"95a49175-2149-4cd4-88fc-e25ca5466ff0\") " pod="openstack/nova-api-badc-account-create-update-njdjh" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.327552 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6stsb\" (UniqueName: \"kubernetes.io/projected/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-kube-api-access-6stsb\") pod \"nova-cell0-db-create-zxg5j\" (UID: \"7ce3702d-5de0-469b-a8e8-b531c4ea5afa\") " pod="openstack/nova-cell0-db-create-zxg5j" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.328610 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-operator-scripts\") pod \"nova-cell0-db-create-zxg5j\" (UID: \"7ce3702d-5de0-469b-a8e8-b531c4ea5afa\") " pod="openstack/nova-cell0-db-create-zxg5j" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.342348 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-f5zbb"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.352708 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6stsb\" (UniqueName: \"kubernetes.io/projected/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-kube-api-access-6stsb\") pod \"nova-cell0-db-create-zxg5j\" (UID: \"7ce3702d-5de0-469b-a8e8-b531c4ea5afa\") " pod="openstack/nova-cell0-db-create-zxg5j" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.379896 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lstm6" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.434667 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95a49175-2149-4cd4-88fc-e25ca5466ff0-operator-scripts\") pod \"nova-api-badc-account-create-update-njdjh\" (UID: \"95a49175-2149-4cd4-88fc-e25ca5466ff0\") " pod="openstack/nova-api-badc-account-create-update-njdjh" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.434754 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-operator-scripts\") pod \"nova-cell1-db-create-f5zbb\" (UID: \"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553\") " pod="openstack/nova-cell1-db-create-f5zbb" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.434810 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh5kl\" (UniqueName: \"kubernetes.io/projected/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-kube-api-access-sh5kl\") pod \"nova-cell1-db-create-f5zbb\" (UID: \"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553\") " pod="openstack/nova-cell1-db-create-f5zbb" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.434841 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb8k2\" (UniqueName: \"kubernetes.io/projected/95a49175-2149-4cd4-88fc-e25ca5466ff0-kube-api-access-fb8k2\") pod \"nova-api-badc-account-create-update-njdjh\" (UID: \"95a49175-2149-4cd4-88fc-e25ca5466ff0\") " pod="openstack/nova-api-badc-account-create-update-njdjh" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.437273 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95a49175-2149-4cd4-88fc-e25ca5466ff0-operator-scripts\") pod \"nova-api-badc-account-create-update-njdjh\" (UID: \"95a49175-2149-4cd4-88fc-e25ca5466ff0\") " pod="openstack/nova-api-badc-account-create-update-njdjh" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.437510 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-operator-scripts\") pod \"nova-cell1-db-create-f5zbb\" (UID: \"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553\") " pod="openstack/nova-cell1-db-create-f5zbb" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.456340 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-216c-account-create-update-gvcn7"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.458745 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-216c-account-create-update-gvcn7" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.460740 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.473796 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh5kl\" (UniqueName: \"kubernetes.io/projected/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-kube-api-access-sh5kl\") pod \"nova-cell1-db-create-f5zbb\" (UID: \"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553\") " pod="openstack/nova-cell1-db-create-f5zbb" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.477399 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb8k2\" (UniqueName: \"kubernetes.io/projected/95a49175-2149-4cd4-88fc-e25ca5466ff0-kube-api-access-fb8k2\") pod \"nova-api-badc-account-create-update-njdjh\" (UID: \"95a49175-2149-4cd4-88fc-e25ca5466ff0\") " pod="openstack/nova-api-badc-account-create-update-njdjh" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.481149 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zxg5j" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.482981 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-216c-account-create-update-gvcn7"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.536704 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tbz8\" (UniqueName: \"kubernetes.io/projected/ec7d33a0-15af-4bec-b44a-a47263f64043-kube-api-access-2tbz8\") pod \"nova-cell0-216c-account-create-update-gvcn7\" (UID: \"ec7d33a0-15af-4bec-b44a-a47263f64043\") " pod="openstack/nova-cell0-216c-account-create-update-gvcn7" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.537083 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec7d33a0-15af-4bec-b44a-a47263f64043-operator-scripts\") pod \"nova-cell0-216c-account-create-update-gvcn7\" (UID: \"ec7d33a0-15af-4bec-b44a-a47263f64043\") " pod="openstack/nova-cell0-216c-account-create-update-gvcn7" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.615663 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-badc-account-create-update-njdjh" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.634832 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-f5zbb" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.639890 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tbz8\" (UniqueName: \"kubernetes.io/projected/ec7d33a0-15af-4bec-b44a-a47263f64043-kube-api-access-2tbz8\") pod \"nova-cell0-216c-account-create-update-gvcn7\" (UID: \"ec7d33a0-15af-4bec-b44a-a47263f64043\") " pod="openstack/nova-cell0-216c-account-create-update-gvcn7" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.639966 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec7d33a0-15af-4bec-b44a-a47263f64043-operator-scripts\") pod \"nova-cell0-216c-account-create-update-gvcn7\" (UID: \"ec7d33a0-15af-4bec-b44a-a47263f64043\") " pod="openstack/nova-cell0-216c-account-create-update-gvcn7" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.646309 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec7d33a0-15af-4bec-b44a-a47263f64043-operator-scripts\") pod \"nova-cell0-216c-account-create-update-gvcn7\" (UID: \"ec7d33a0-15af-4bec-b44a-a47263f64043\") " pod="openstack/nova-cell0-216c-account-create-update-gvcn7" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.675106 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-0136-account-create-update-4q6m4"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.678225 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0136-account-create-update-4q6m4" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.682963 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.685544 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tbz8\" (UniqueName: \"kubernetes.io/projected/ec7d33a0-15af-4bec-b44a-a47263f64043-kube-api-access-2tbz8\") pod \"nova-cell0-216c-account-create-update-gvcn7\" (UID: \"ec7d33a0-15af-4bec-b44a-a47263f64043\") " pod="openstack/nova-cell0-216c-account-create-update-gvcn7" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.751167 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0136-account-create-update-4q6m4"] Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.856418 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b017666f-1224-49c2-842c-d2e0034556e7-operator-scripts\") pod \"nova-cell1-0136-account-create-update-4q6m4\" (UID: \"b017666f-1224-49c2-842c-d2e0034556e7\") " pod="openstack/nova-cell1-0136-account-create-update-4q6m4" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.857573 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtgl4\" (UniqueName: \"kubernetes.io/projected/b017666f-1224-49c2-842c-d2e0034556e7-kube-api-access-gtgl4\") pod \"nova-cell1-0136-account-create-update-4q6m4\" (UID: \"b017666f-1224-49c2-842c-d2e0034556e7\") " pod="openstack/nova-cell1-0136-account-create-update-4q6m4" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.957104 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-216c-account-create-update-gvcn7" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.961468 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b017666f-1224-49c2-842c-d2e0034556e7-operator-scripts\") pod \"nova-cell1-0136-account-create-update-4q6m4\" (UID: \"b017666f-1224-49c2-842c-d2e0034556e7\") " pod="openstack/nova-cell1-0136-account-create-update-4q6m4" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.961950 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtgl4\" (UniqueName: \"kubernetes.io/projected/b017666f-1224-49c2-842c-d2e0034556e7-kube-api-access-gtgl4\") pod \"nova-cell1-0136-account-create-update-4q6m4\" (UID: \"b017666f-1224-49c2-842c-d2e0034556e7\") " pod="openstack/nova-cell1-0136-account-create-update-4q6m4" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.969400 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b017666f-1224-49c2-842c-d2e0034556e7-operator-scripts\") pod \"nova-cell1-0136-account-create-update-4q6m4\" (UID: \"b017666f-1224-49c2-842c-d2e0034556e7\") " pod="openstack/nova-cell1-0136-account-create-update-4q6m4" Dec 06 07:18:08 crc kubenswrapper[4945]: I1206 07:18:08.989046 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtgl4\" (UniqueName: \"kubernetes.io/projected/b017666f-1224-49c2-842c-d2e0034556e7-kube-api-access-gtgl4\") pod \"nova-cell1-0136-account-create-update-4q6m4\" (UID: \"b017666f-1224-49c2-842c-d2e0034556e7\") " pod="openstack/nova-cell1-0136-account-create-update-4q6m4" Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.018830 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0136-account-create-update-4q6m4" Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.137268 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lstm6"] Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.365578 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-zxg5j"] Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.550356 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-badc-account-create-update-njdjh"] Dec 06 07:18:09 crc kubenswrapper[4945]: W1206 07:18:09.576720 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95a49175_2149_4cd4_88fc_e25ca5466ff0.slice/crio-1f03681aee7176b4bfb1e4e58f25da126c0207455dfcfdffdb57910907e6ee83 WatchSource:0}: Error finding container 1f03681aee7176b4bfb1e4e58f25da126c0207455dfcfdffdb57910907e6ee83: Status 404 returned error can't find the container with id 1f03681aee7176b4bfb1e4e58f25da126c0207455dfcfdffdb57910907e6ee83 Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.602195 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lstm6" event={"ID":"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd","Type":"ContainerStarted","Data":"fcba1c2dcac5da3efdccd211f37d3778051860206a95c83a23d221f05b17017f"} Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.606645 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.606667 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.608643 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zxg5j" event={"ID":"7ce3702d-5de0-469b-a8e8-b531c4ea5afa","Type":"ContainerStarted","Data":"f0e4adb0f8845f8ab6b6dddd79ae098da45706d447859ab62d5d52e6ae4e50d3"} Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.772532 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-f5zbb"] Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.965199 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0136-account-create-update-4q6m4"] Dec 06 07:18:09 crc kubenswrapper[4945]: I1206 07:18:09.991196 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-216c-account-create-update-gvcn7"] Dec 06 07:18:10 crc kubenswrapper[4945]: W1206 07:18:10.008349 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec7d33a0_15af_4bec_b44a_a47263f64043.slice/crio-49a7bfbbecdaf1bf5bde546afab00013b6346fe79ba38fc3738bc6e4c663dcd6 WatchSource:0}: Error finding container 49a7bfbbecdaf1bf5bde546afab00013b6346fe79ba38fc3738bc6e4c663dcd6: Status 404 returned error can't find the container with id 49a7bfbbecdaf1bf5bde546afab00013b6346fe79ba38fc3738bc6e4c663dcd6 Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.652355 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-badc-account-create-update-njdjh" event={"ID":"95a49175-2149-4cd4-88fc-e25ca5466ff0","Type":"ContainerStarted","Data":"c13f81a9e25d94e04593bed44e1a89fc9b9c9767b12d18a9ce9a454c2b1fb953"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.652757 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-badc-account-create-update-njdjh" event={"ID":"95a49175-2149-4cd4-88fc-e25ca5466ff0","Type":"ContainerStarted","Data":"1f03681aee7176b4bfb1e4e58f25da126c0207455dfcfdffdb57910907e6ee83"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.678773 4945 generic.go:334] "Generic (PLEG): container finished" podID="0f52cc49-cbae-487e-8d29-bdcb20fa6ddd" containerID="87cd0b4a674d5d3593011da803e898cb217a594e41b955bcb125ea3f0c1b9ff8" exitCode=0 Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.678929 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lstm6" event={"ID":"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd","Type":"ContainerDied","Data":"87cd0b4a674d5d3593011da803e898cb217a594e41b955bcb125ea3f0c1b9ff8"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.702740 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0136-account-create-update-4q6m4" event={"ID":"b017666f-1224-49c2-842c-d2e0034556e7","Type":"ContainerStarted","Data":"acd8fb52b675c118e310b60c8ee78c1ae8f429af506bdaf70f33be2b2e7d0f4d"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.702810 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0136-account-create-update-4q6m4" event={"ID":"b017666f-1224-49c2-842c-d2e0034556e7","Type":"ContainerStarted","Data":"138c3ab56e73c2c8befb4e24606c7a637b313e0feaf083f959efb3fc47d18959"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.726014 4945 generic.go:334] "Generic (PLEG): container finished" podID="7ce3702d-5de0-469b-a8e8-b531c4ea5afa" containerID="ab370550567ae529abd1c5be85f05c4c95e05cb15eb9877d9089c7f9cdaec299" exitCode=0 Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.726129 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zxg5j" event={"ID":"7ce3702d-5de0-469b-a8e8-b531c4ea5afa","Type":"ContainerDied","Data":"ab370550567ae529abd1c5be85f05c4c95e05cb15eb9877d9089c7f9cdaec299"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.735180 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-badc-account-create-update-njdjh" podStartSLOduration=2.7351312759999997 podStartE2EDuration="2.735131276s" podCreationTimestamp="2025-12-06 07:18:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:18:10.702117233 +0000 UTC m=+1504.156978297" watchObservedRunningTime="2025-12-06 07:18:10.735131276 +0000 UTC m=+1504.189992330" Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.735506 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-f5zbb" event={"ID":"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553","Type":"ContainerStarted","Data":"36fc5adc9c35a27308a897b987a0ead13d775ca1f0ea63ba9ab9e64e112df411"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.735569 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-f5zbb" event={"ID":"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553","Type":"ContainerStarted","Data":"3eefcc8961a072818ebc19d7df5878d8ed9b78583bbf47bb0d5843a149c9ffd3"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.742808 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-216c-account-create-update-gvcn7" event={"ID":"ec7d33a0-15af-4bec-b44a-a47263f64043","Type":"ContainerStarted","Data":"7cb7cbad5c146d1fa5c2816193c03ecbbc2c2e83f7d5dbab3b49095c873c3919"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.742882 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-216c-account-create-update-gvcn7" event={"ID":"ec7d33a0-15af-4bec-b44a-a47263f64043","Type":"ContainerStarted","Data":"49a7bfbbecdaf1bf5bde546afab00013b6346fe79ba38fc3738bc6e4c663dcd6"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.757547 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57333dda-f06f-4d53-86ee-8514ff7151ea","Type":"ContainerStarted","Data":"810c251001ae0fdd0e5c933600cc3f650cadee55ff768f7deb5c3d7f87156cf7"} Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.758617 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.763613 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-0136-account-create-update-4q6m4" podStartSLOduration=2.7635852659999998 podStartE2EDuration="2.763585266s" podCreationTimestamp="2025-12-06 07:18:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:18:10.749552347 +0000 UTC m=+1504.204413391" watchObservedRunningTime="2025-12-06 07:18:10.763585266 +0000 UTC m=+1504.218446310" Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.829866 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-f5zbb" podStartSLOduration=2.8298389090000002 podStartE2EDuration="2.829838909s" podCreationTimestamp="2025-12-06 07:18:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:18:10.790397222 +0000 UTC m=+1504.245258266" watchObservedRunningTime="2025-12-06 07:18:10.829838909 +0000 UTC m=+1504.284699953" Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.836084 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.003028859 podStartE2EDuration="6.836061617s" podCreationTimestamp="2025-12-06 07:18:04 +0000 UTC" firstStartedPulling="2025-12-06 07:18:05.308846085 +0000 UTC m=+1498.763707129" lastFinishedPulling="2025-12-06 07:18:09.141878843 +0000 UTC m=+1502.596739887" observedRunningTime="2025-12-06 07:18:10.817777162 +0000 UTC m=+1504.272638206" watchObservedRunningTime="2025-12-06 07:18:10.836061617 +0000 UTC m=+1504.290922661" Dec 06 07:18:10 crc kubenswrapper[4945]: I1206 07:18:10.848061 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-216c-account-create-update-gvcn7" podStartSLOduration=2.848023501 podStartE2EDuration="2.848023501s" podCreationTimestamp="2025-12-06 07:18:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:18:10.846106529 +0000 UTC m=+1504.300967573" watchObservedRunningTime="2025-12-06 07:18:10.848023501 +0000 UTC m=+1504.302884545" Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.132053 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.132236 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.134836 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.189760 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.190174 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.317991 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.771256 4945 generic.go:334] "Generic (PLEG): container finished" podID="ec7d33a0-15af-4bec-b44a-a47263f64043" containerID="7cb7cbad5c146d1fa5c2816193c03ecbbc2c2e83f7d5dbab3b49095c873c3919" exitCode=0 Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.771335 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-216c-account-create-update-gvcn7" event={"ID":"ec7d33a0-15af-4bec-b44a-a47263f64043","Type":"ContainerDied","Data":"7cb7cbad5c146d1fa5c2816193c03ecbbc2c2e83f7d5dbab3b49095c873c3919"} Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.774413 4945 generic.go:334] "Generic (PLEG): container finished" podID="95a49175-2149-4cd4-88fc-e25ca5466ff0" containerID="c13f81a9e25d94e04593bed44e1a89fc9b9c9767b12d18a9ce9a454c2b1fb953" exitCode=0 Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.774501 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-badc-account-create-update-njdjh" event={"ID":"95a49175-2149-4cd4-88fc-e25ca5466ff0","Type":"ContainerDied","Data":"c13f81a9e25d94e04593bed44e1a89fc9b9c9767b12d18a9ce9a454c2b1fb953"} Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.777521 4945 generic.go:334] "Generic (PLEG): container finished" podID="b017666f-1224-49c2-842c-d2e0034556e7" containerID="acd8fb52b675c118e310b60c8ee78c1ae8f429af506bdaf70f33be2b2e7d0f4d" exitCode=0 Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.777615 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0136-account-create-update-4q6m4" event={"ID":"b017666f-1224-49c2-842c-d2e0034556e7","Type":"ContainerDied","Data":"acd8fb52b675c118e310b60c8ee78c1ae8f429af506bdaf70f33be2b2e7d0f4d"} Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.780527 4945 generic.go:334] "Generic (PLEG): container finished" podID="83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553" containerID="36fc5adc9c35a27308a897b987a0ead13d775ca1f0ea63ba9ab9e64e112df411" exitCode=0 Dec 06 07:18:11 crc kubenswrapper[4945]: I1206 07:18:11.781757 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-f5zbb" event={"ID":"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553","Type":"ContainerDied","Data":"36fc5adc9c35a27308a897b987a0ead13d775ca1f0ea63ba9ab9e64e112df411"} Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.045300 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.329310 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zxg5j" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.335465 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lstm6" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.375671 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-operator-scripts\") pod \"7ce3702d-5de0-469b-a8e8-b531c4ea5afa\" (UID: \"7ce3702d-5de0-469b-a8e8-b531c4ea5afa\") " Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.376052 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6stsb\" (UniqueName: \"kubernetes.io/projected/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-kube-api-access-6stsb\") pod \"7ce3702d-5de0-469b-a8e8-b531c4ea5afa\" (UID: \"7ce3702d-5de0-469b-a8e8-b531c4ea5afa\") " Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.376271 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ce3702d-5de0-469b-a8e8-b531c4ea5afa" (UID: "7ce3702d-5de0-469b-a8e8-b531c4ea5afa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.377465 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.384210 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-kube-api-access-6stsb" (OuterVolumeSpecName: "kube-api-access-6stsb") pod "7ce3702d-5de0-469b-a8e8-b531c4ea5afa" (UID: "7ce3702d-5de0-469b-a8e8-b531c4ea5afa"). InnerVolumeSpecName "kube-api-access-6stsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.479650 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thkf6\" (UniqueName: \"kubernetes.io/projected/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-kube-api-access-thkf6\") pod \"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd\" (UID: \"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd\") " Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.480009 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-operator-scripts\") pod \"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd\" (UID: \"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd\") " Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.481092 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6stsb\" (UniqueName: \"kubernetes.io/projected/7ce3702d-5de0-469b-a8e8-b531c4ea5afa-kube-api-access-6stsb\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.481850 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f52cc49-cbae-487e-8d29-bdcb20fa6ddd" (UID: "0f52cc49-cbae-487e-8d29-bdcb20fa6ddd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.487534 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-kube-api-access-thkf6" (OuterVolumeSpecName: "kube-api-access-thkf6") pod "0f52cc49-cbae-487e-8d29-bdcb20fa6ddd" (UID: "0f52cc49-cbae-487e-8d29-bdcb20fa6ddd"). InnerVolumeSpecName "kube-api-access-thkf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.583074 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.583123 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thkf6\" (UniqueName: \"kubernetes.io/projected/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd-kube-api-access-thkf6\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.695021 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.793360 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lstm6" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.793346 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lstm6" event={"ID":"0f52cc49-cbae-487e-8d29-bdcb20fa6ddd","Type":"ContainerDied","Data":"fcba1c2dcac5da3efdccd211f37d3778051860206a95c83a23d221f05b17017f"} Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.793516 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcba1c2dcac5da3efdccd211f37d3778051860206a95c83a23d221f05b17017f" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.795723 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zxg5j" Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.808396 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zxg5j" event={"ID":"7ce3702d-5de0-469b-a8e8-b531c4ea5afa","Type":"ContainerDied","Data":"f0e4adb0f8845f8ab6b6dddd79ae098da45706d447859ab62d5d52e6ae4e50d3"} Dec 06 07:18:12 crc kubenswrapper[4945]: I1206 07:18:12.808480 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0e4adb0f8845f8ab6b6dddd79ae098da45706d447859ab62d5d52e6ae4e50d3" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.377550 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-216c-account-create-update-gvcn7" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.516386 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec7d33a0-15af-4bec-b44a-a47263f64043-operator-scripts\") pod \"ec7d33a0-15af-4bec-b44a-a47263f64043\" (UID: \"ec7d33a0-15af-4bec-b44a-a47263f64043\") " Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.516522 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tbz8\" (UniqueName: \"kubernetes.io/projected/ec7d33a0-15af-4bec-b44a-a47263f64043-kube-api-access-2tbz8\") pod \"ec7d33a0-15af-4bec-b44a-a47263f64043\" (UID: \"ec7d33a0-15af-4bec-b44a-a47263f64043\") " Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.518265 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec7d33a0-15af-4bec-b44a-a47263f64043-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec7d33a0-15af-4bec-b44a-a47263f64043" (UID: "ec7d33a0-15af-4bec-b44a-a47263f64043"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.522499 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec7d33a0-15af-4bec-b44a-a47263f64043-kube-api-access-2tbz8" (OuterVolumeSpecName: "kube-api-access-2tbz8") pod "ec7d33a0-15af-4bec-b44a-a47263f64043" (UID: "ec7d33a0-15af-4bec-b44a-a47263f64043"). InnerVolumeSpecName "kube-api-access-2tbz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.583535 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-badc-account-create-update-njdjh" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.590811 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0136-account-create-update-4q6m4" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.622037 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tbz8\" (UniqueName: \"kubernetes.io/projected/ec7d33a0-15af-4bec-b44a-a47263f64043-kube-api-access-2tbz8\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.622080 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec7d33a0-15af-4bec-b44a-a47263f64043-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.624730 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-f5zbb" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.723103 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtgl4\" (UniqueName: \"kubernetes.io/projected/b017666f-1224-49c2-842c-d2e0034556e7-kube-api-access-gtgl4\") pod \"b017666f-1224-49c2-842c-d2e0034556e7\" (UID: \"b017666f-1224-49c2-842c-d2e0034556e7\") " Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.723337 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b017666f-1224-49c2-842c-d2e0034556e7-operator-scripts\") pod \"b017666f-1224-49c2-842c-d2e0034556e7\" (UID: \"b017666f-1224-49c2-842c-d2e0034556e7\") " Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.723898 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b017666f-1224-49c2-842c-d2e0034556e7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b017666f-1224-49c2-842c-d2e0034556e7" (UID: "b017666f-1224-49c2-842c-d2e0034556e7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.723970 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-operator-scripts\") pod \"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553\" (UID: \"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553\") " Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.724444 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553" (UID: "83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.724504 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95a49175-2149-4cd4-88fc-e25ca5466ff0-operator-scripts\") pod \"95a49175-2149-4cd4-88fc-e25ca5466ff0\" (UID: \"95a49175-2149-4cd4-88fc-e25ca5466ff0\") " Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.724556 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh5kl\" (UniqueName: \"kubernetes.io/projected/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-kube-api-access-sh5kl\") pod \"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553\" (UID: \"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553\") " Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.724598 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb8k2\" (UniqueName: \"kubernetes.io/projected/95a49175-2149-4cd4-88fc-e25ca5466ff0-kube-api-access-fb8k2\") pod \"95a49175-2149-4cd4-88fc-e25ca5466ff0\" (UID: \"95a49175-2149-4cd4-88fc-e25ca5466ff0\") " Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.725159 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b017666f-1224-49c2-842c-d2e0034556e7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.725177 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.726661 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95a49175-2149-4cd4-88fc-e25ca5466ff0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95a49175-2149-4cd4-88fc-e25ca5466ff0" (UID: "95a49175-2149-4cd4-88fc-e25ca5466ff0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.727158 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b017666f-1224-49c2-842c-d2e0034556e7-kube-api-access-gtgl4" (OuterVolumeSpecName: "kube-api-access-gtgl4") pod "b017666f-1224-49c2-842c-d2e0034556e7" (UID: "b017666f-1224-49c2-842c-d2e0034556e7"). InnerVolumeSpecName "kube-api-access-gtgl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.729242 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a49175-2149-4cd4-88fc-e25ca5466ff0-kube-api-access-fb8k2" (OuterVolumeSpecName: "kube-api-access-fb8k2") pod "95a49175-2149-4cd4-88fc-e25ca5466ff0" (UID: "95a49175-2149-4cd4-88fc-e25ca5466ff0"). InnerVolumeSpecName "kube-api-access-fb8k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.739502 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-kube-api-access-sh5kl" (OuterVolumeSpecName: "kube-api-access-sh5kl") pod "83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553" (UID: "83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553"). InnerVolumeSpecName "kube-api-access-sh5kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.826811 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtgl4\" (UniqueName: \"kubernetes.io/projected/b017666f-1224-49c2-842c-d2e0034556e7-kube-api-access-gtgl4\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.826845 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95a49175-2149-4cd4-88fc-e25ca5466ff0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.826856 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh5kl\" (UniqueName: \"kubernetes.io/projected/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553-kube-api-access-sh5kl\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.826868 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb8k2\" (UniqueName: \"kubernetes.io/projected/95a49175-2149-4cd4-88fc-e25ca5466ff0-kube-api-access-fb8k2\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.827167 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-f5zbb" event={"ID":"83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553","Type":"ContainerDied","Data":"3eefcc8961a072818ebc19d7df5878d8ed9b78583bbf47bb0d5843a149c9ffd3"} Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.827200 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3eefcc8961a072818ebc19d7df5878d8ed9b78583bbf47bb0d5843a149c9ffd3" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.827261 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-f5zbb" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.838764 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-216c-account-create-update-gvcn7" event={"ID":"ec7d33a0-15af-4bec-b44a-a47263f64043","Type":"ContainerDied","Data":"49a7bfbbecdaf1bf5bde546afab00013b6346fe79ba38fc3738bc6e4c663dcd6"} Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.838817 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49a7bfbbecdaf1bf5bde546afab00013b6346fe79ba38fc3738bc6e4c663dcd6" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.838877 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-216c-account-create-update-gvcn7" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.861739 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-badc-account-create-update-njdjh" event={"ID":"95a49175-2149-4cd4-88fc-e25ca5466ff0","Type":"ContainerDied","Data":"1f03681aee7176b4bfb1e4e58f25da126c0207455dfcfdffdb57910907e6ee83"} Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.861799 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f03681aee7176b4bfb1e4e58f25da126c0207455dfcfdffdb57910907e6ee83" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.861895 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-badc-account-create-update-njdjh" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.883801 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0136-account-create-update-4q6m4" event={"ID":"b017666f-1224-49c2-842c-d2e0034556e7","Type":"ContainerDied","Data":"138c3ab56e73c2c8befb4e24606c7a637b313e0feaf083f959efb3fc47d18959"} Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.883860 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="138c3ab56e73c2c8befb4e24606c7a637b313e0feaf083f959efb3fc47d18959" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.883867 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0136-account-create-update-4q6m4" Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.884011 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="ceilometer-central-agent" containerID="cri-o://024ef6e2c14bea89f70ebb385437b4ea6f8705e27f84943de8f3ea92ab4132e1" gracePeriod=30 Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.884363 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="proxy-httpd" containerID="cri-o://810c251001ae0fdd0e5c933600cc3f650cadee55ff768f7deb5c3d7f87156cf7" gracePeriod=30 Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.884273 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="sg-core" containerID="cri-o://bbfb74902890e9a61b24d3f56f0e7a8bb54514b8af03f6d9cb65ceac20b20159" gracePeriod=30 Dec 06 07:18:13 crc kubenswrapper[4945]: I1206 07:18:13.884408 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="ceilometer-notification-agent" containerID="cri-o://cfacc27b96dd10b0742f177f8e6346a834f8709ab3a4aea54db05a0c7502f042" gracePeriod=30 Dec 06 07:18:14 crc kubenswrapper[4945]: I1206 07:18:14.913356 4945 generic.go:334] "Generic (PLEG): container finished" podID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerID="810c251001ae0fdd0e5c933600cc3f650cadee55ff768f7deb5c3d7f87156cf7" exitCode=0 Dec 06 07:18:14 crc kubenswrapper[4945]: I1206 07:18:14.915368 4945 generic.go:334] "Generic (PLEG): container finished" podID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerID="bbfb74902890e9a61b24d3f56f0e7a8bb54514b8af03f6d9cb65ceac20b20159" exitCode=2 Dec 06 07:18:14 crc kubenswrapper[4945]: I1206 07:18:14.915385 4945 generic.go:334] "Generic (PLEG): container finished" podID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerID="cfacc27b96dd10b0742f177f8e6346a834f8709ab3a4aea54db05a0c7502f042" exitCode=0 Dec 06 07:18:14 crc kubenswrapper[4945]: I1206 07:18:14.915445 4945 generic.go:334] "Generic (PLEG): container finished" podID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerID="024ef6e2c14bea89f70ebb385437b4ea6f8705e27f84943de8f3ea92ab4132e1" exitCode=0 Dec 06 07:18:14 crc kubenswrapper[4945]: I1206 07:18:14.913419 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57333dda-f06f-4d53-86ee-8514ff7151ea","Type":"ContainerDied","Data":"810c251001ae0fdd0e5c933600cc3f650cadee55ff768f7deb5c3d7f87156cf7"} Dec 06 07:18:14 crc kubenswrapper[4945]: I1206 07:18:14.915530 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57333dda-f06f-4d53-86ee-8514ff7151ea","Type":"ContainerDied","Data":"bbfb74902890e9a61b24d3f56f0e7a8bb54514b8af03f6d9cb65ceac20b20159"} Dec 06 07:18:14 crc kubenswrapper[4945]: I1206 07:18:14.915555 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57333dda-f06f-4d53-86ee-8514ff7151ea","Type":"ContainerDied","Data":"cfacc27b96dd10b0742f177f8e6346a834f8709ab3a4aea54db05a0c7502f042"} Dec 06 07:18:14 crc kubenswrapper[4945]: I1206 07:18:14.915575 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57333dda-f06f-4d53-86ee-8514ff7151ea","Type":"ContainerDied","Data":"024ef6e2c14bea89f70ebb385437b4ea6f8705e27f84943de8f3ea92ab4132e1"} Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.045469 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.164102 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-log-httpd\") pod \"57333dda-f06f-4d53-86ee-8514ff7151ea\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.164153 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-sg-core-conf-yaml\") pod \"57333dda-f06f-4d53-86ee-8514ff7151ea\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.164299 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-config-data\") pod \"57333dda-f06f-4d53-86ee-8514ff7151ea\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.164347 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-scripts\") pod \"57333dda-f06f-4d53-86ee-8514ff7151ea\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.164420 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-425g7\" (UniqueName: \"kubernetes.io/projected/57333dda-f06f-4d53-86ee-8514ff7151ea-kube-api-access-425g7\") pod \"57333dda-f06f-4d53-86ee-8514ff7151ea\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.164446 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-run-httpd\") pod \"57333dda-f06f-4d53-86ee-8514ff7151ea\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.164497 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-combined-ca-bundle\") pod \"57333dda-f06f-4d53-86ee-8514ff7151ea\" (UID: \"57333dda-f06f-4d53-86ee-8514ff7151ea\") " Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.166554 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "57333dda-f06f-4d53-86ee-8514ff7151ea" (UID: "57333dda-f06f-4d53-86ee-8514ff7151ea"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.166881 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "57333dda-f06f-4d53-86ee-8514ff7151ea" (UID: "57333dda-f06f-4d53-86ee-8514ff7151ea"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.172567 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-scripts" (OuterVolumeSpecName: "scripts") pod "57333dda-f06f-4d53-86ee-8514ff7151ea" (UID: "57333dda-f06f-4d53-86ee-8514ff7151ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.172989 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57333dda-f06f-4d53-86ee-8514ff7151ea-kube-api-access-425g7" (OuterVolumeSpecName: "kube-api-access-425g7") pod "57333dda-f06f-4d53-86ee-8514ff7151ea" (UID: "57333dda-f06f-4d53-86ee-8514ff7151ea"). InnerVolumeSpecName "kube-api-access-425g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.198162 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "57333dda-f06f-4d53-86ee-8514ff7151ea" (UID: "57333dda-f06f-4d53-86ee-8514ff7151ea"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.261536 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57333dda-f06f-4d53-86ee-8514ff7151ea" (UID: "57333dda-f06f-4d53-86ee-8514ff7151ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.267047 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.267219 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.267321 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.267459 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-425g7\" (UniqueName: \"kubernetes.io/projected/57333dda-f06f-4d53-86ee-8514ff7151ea-kube-api-access-425g7\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.267566 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57333dda-f06f-4d53-86ee-8514ff7151ea-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.267653 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.288362 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-config-data" (OuterVolumeSpecName: "config-data") pod "57333dda-f06f-4d53-86ee-8514ff7151ea" (UID: "57333dda-f06f-4d53-86ee-8514ff7151ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.370169 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57333dda-f06f-4d53-86ee-8514ff7151ea-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.927573 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57333dda-f06f-4d53-86ee-8514ff7151ea","Type":"ContainerDied","Data":"9bb626a8ff6b93dc38fef4868ad301c7fb337a8953d150c0d30d0f318506671d"} Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.927942 4945 scope.go:117] "RemoveContainer" containerID="810c251001ae0fdd0e5c933600cc3f650cadee55ff768f7deb5c3d7f87156cf7" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.927633 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.952302 4945 scope.go:117] "RemoveContainer" containerID="bbfb74902890e9a61b24d3f56f0e7a8bb54514b8af03f6d9cb65ceac20b20159" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.975455 4945 scope.go:117] "RemoveContainer" containerID="cfacc27b96dd10b0742f177f8e6346a834f8709ab3a4aea54db05a0c7502f042" Dec 06 07:18:15 crc kubenswrapper[4945]: I1206 07:18:15.981013 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.003311 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.009581 4945 scope.go:117] "RemoveContainer" containerID="024ef6e2c14bea89f70ebb385437b4ea6f8705e27f84943de8f3ea92ab4132e1" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.016548 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:16 crc kubenswrapper[4945]: E1206 07:18:16.017190 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b017666f-1224-49c2-842c-d2e0034556e7" containerName="mariadb-account-create-update" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017212 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b017666f-1224-49c2-842c-d2e0034556e7" containerName="mariadb-account-create-update" Dec 06 07:18:16 crc kubenswrapper[4945]: E1206 07:18:16.017233 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="ceilometer-central-agent" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017242 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="ceilometer-central-agent" Dec 06 07:18:16 crc kubenswrapper[4945]: E1206 07:18:16.017267 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553" containerName="mariadb-database-create" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017291 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553" containerName="mariadb-database-create" Dec 06 07:18:16 crc kubenswrapper[4945]: E1206 07:18:16.017303 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce3702d-5de0-469b-a8e8-b531c4ea5afa" containerName="mariadb-database-create" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017309 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce3702d-5de0-469b-a8e8-b531c4ea5afa" containerName="mariadb-database-create" Dec 06 07:18:16 crc kubenswrapper[4945]: E1206 07:18:16.017321 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="proxy-httpd" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017328 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="proxy-httpd" Dec 06 07:18:16 crc kubenswrapper[4945]: E1206 07:18:16.017350 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="ceilometer-notification-agent" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017359 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="ceilometer-notification-agent" Dec 06 07:18:16 crc kubenswrapper[4945]: E1206 07:18:16.017370 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="sg-core" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017377 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="sg-core" Dec 06 07:18:16 crc kubenswrapper[4945]: E1206 07:18:16.017389 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a49175-2149-4cd4-88fc-e25ca5466ff0" containerName="mariadb-account-create-update" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017396 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a49175-2149-4cd4-88fc-e25ca5466ff0" containerName="mariadb-account-create-update" Dec 06 07:18:16 crc kubenswrapper[4945]: E1206 07:18:16.017410 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f52cc49-cbae-487e-8d29-bdcb20fa6ddd" containerName="mariadb-database-create" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017416 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f52cc49-cbae-487e-8d29-bdcb20fa6ddd" containerName="mariadb-database-create" Dec 06 07:18:16 crc kubenswrapper[4945]: E1206 07:18:16.017435 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec7d33a0-15af-4bec-b44a-a47263f64043" containerName="mariadb-account-create-update" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017441 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec7d33a0-15af-4bec-b44a-a47263f64043" containerName="mariadb-account-create-update" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017628 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="proxy-httpd" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017644 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b017666f-1224-49c2-842c-d2e0034556e7" containerName="mariadb-account-create-update" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017663 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553" containerName="mariadb-database-create" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017679 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec7d33a0-15af-4bec-b44a-a47263f64043" containerName="mariadb-account-create-update" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017699 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce3702d-5de0-469b-a8e8-b531c4ea5afa" containerName="mariadb-database-create" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017715 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f52cc49-cbae-487e-8d29-bdcb20fa6ddd" containerName="mariadb-database-create" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017729 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="ceilometer-central-agent" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017743 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="ceilometer-notification-agent" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017755 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a49175-2149-4cd4-88fc-e25ca5466ff0" containerName="mariadb-account-create-update" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.017769 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" containerName="sg-core" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.022950 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.026137 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.033934 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.034011 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.083150 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.083223 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-log-httpd\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.083293 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-scripts\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.083413 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-config-data\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.083511 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsmvv\" (UniqueName: \"kubernetes.io/projected/72d55594-ed65-4a69-b8cf-1a8925ea32a6-kube-api-access-vsmvv\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.083544 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.083590 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-run-httpd\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.185759 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsmvv\" (UniqueName: \"kubernetes.io/projected/72d55594-ed65-4a69-b8cf-1a8925ea32a6-kube-api-access-vsmvv\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.186175 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.186274 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-run-httpd\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.186411 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.186525 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-log-httpd\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.186637 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-scripts\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.186813 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-config-data\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.187866 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-run-httpd\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.188053 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-log-httpd\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.193238 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-scripts\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.193669 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.193953 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.206842 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-config-data\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.215529 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsmvv\" (UniqueName: \"kubernetes.io/projected/72d55594-ed65-4a69-b8cf-1a8925ea32a6-kube-api-access-vsmvv\") pod \"ceilometer-0\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.358149 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.862700 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.943400 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72d55594-ed65-4a69-b8cf-1a8925ea32a6","Type":"ContainerStarted","Data":"1fe9a5d504e096c96da60c21bcff141854a6ffeec4801aeef7693ee44aa93a75"} Dec 06 07:18:16 crc kubenswrapper[4945]: I1206 07:18:16.985153 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57333dda-f06f-4d53-86ee-8514ff7151ea" path="/var/lib/kubelet/pods/57333dda-f06f-4d53-86ee-8514ff7151ea/volumes" Dec 06 07:18:17 crc kubenswrapper[4945]: I1206 07:18:17.956849 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72d55594-ed65-4a69-b8cf-1a8925ea32a6","Type":"ContainerStarted","Data":"cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1"} Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.709305 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zljkx"] Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.712112 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.716677 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-5tvtw" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.716824 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.717066 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.727790 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zljkx"] Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.852986 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-scripts\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.853099 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.853140 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg8hp\" (UniqueName: \"kubernetes.io/projected/72cf544a-1950-42b5-b244-b20339a8c702-kube-api-access-gg8hp\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.853201 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-config-data\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.869397 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="72e9a7b2-b482-4130-819a-95ac1106e218" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.954644 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-config-data\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.954730 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-scripts\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.954802 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.954837 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg8hp\" (UniqueName: \"kubernetes.io/projected/72cf544a-1950-42b5-b244-b20339a8c702-kube-api-access-gg8hp\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.962158 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.970258 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-config-data\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.979163 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-scripts\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:18 crc kubenswrapper[4945]: I1206 07:18:18.985755 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg8hp\" (UniqueName: \"kubernetes.io/projected/72cf544a-1950-42b5-b244-b20339a8c702-kube-api-access-gg8hp\") pod \"nova-cell0-conductor-db-sync-zljkx\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:19 crc kubenswrapper[4945]: I1206 07:18:19.002322 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72d55594-ed65-4a69-b8cf-1a8925ea32a6","Type":"ContainerStarted","Data":"92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065"} Dec 06 07:18:19 crc kubenswrapper[4945]: I1206 07:18:19.040304 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:19 crc kubenswrapper[4945]: I1206 07:18:19.964588 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zljkx"] Dec 06 07:18:19 crc kubenswrapper[4945]: W1206 07:18:19.973372 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72cf544a_1950_42b5_b244_b20339a8c702.slice/crio-eb292f3be12d664d670eea7c64176c074d16f5437ba2752cf434b73034d0fe6f WatchSource:0}: Error finding container eb292f3be12d664d670eea7c64176c074d16f5437ba2752cf434b73034d0fe6f: Status 404 returned error can't find the container with id eb292f3be12d664d670eea7c64176c074d16f5437ba2752cf434b73034d0fe6f Dec 06 07:18:20 crc kubenswrapper[4945]: I1206 07:18:20.016901 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72d55594-ed65-4a69-b8cf-1a8925ea32a6","Type":"ContainerStarted","Data":"b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6"} Dec 06 07:18:20 crc kubenswrapper[4945]: I1206 07:18:20.019734 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zljkx" event={"ID":"72cf544a-1950-42b5-b244-b20339a8c702","Type":"ContainerStarted","Data":"eb292f3be12d664d670eea7c64176c074d16f5437ba2752cf434b73034d0fe6f"} Dec 06 07:18:22 crc kubenswrapper[4945]: I1206 07:18:22.048080 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72d55594-ed65-4a69-b8cf-1a8925ea32a6","Type":"ContainerStarted","Data":"133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf"} Dec 06 07:18:22 crc kubenswrapper[4945]: I1206 07:18:22.049167 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 07:18:22 crc kubenswrapper[4945]: I1206 07:18:22.082112 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.6542326469999997 podStartE2EDuration="7.082091247s" podCreationTimestamp="2025-12-06 07:18:15 +0000 UTC" firstStartedPulling="2025-12-06 07:18:16.869204768 +0000 UTC m=+1510.324065812" lastFinishedPulling="2025-12-06 07:18:21.297063368 +0000 UTC m=+1514.751924412" observedRunningTime="2025-12-06 07:18:22.079925048 +0000 UTC m=+1515.534786102" watchObservedRunningTime="2025-12-06 07:18:22.082091247 +0000 UTC m=+1515.536952291" Dec 06 07:18:29 crc kubenswrapper[4945]: I1206 07:18:29.132688 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zljkx" event={"ID":"72cf544a-1950-42b5-b244-b20339a8c702","Type":"ContainerStarted","Data":"724de5f0c3d669660e6698a3c99a4875f3d5e9150ef8166ff6741bfbba23a072"} Dec 06 07:18:29 crc kubenswrapper[4945]: I1206 07:18:29.156803 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-zljkx" podStartSLOduration=2.898568959 podStartE2EDuration="11.156778535s" podCreationTimestamp="2025-12-06 07:18:18 +0000 UTC" firstStartedPulling="2025-12-06 07:18:19.976652797 +0000 UTC m=+1513.431513841" lastFinishedPulling="2025-12-06 07:18:28.234862373 +0000 UTC m=+1521.689723417" observedRunningTime="2025-12-06 07:18:29.151920553 +0000 UTC m=+1522.606781607" watchObservedRunningTime="2025-12-06 07:18:29.156778535 +0000 UTC m=+1522.611639579" Dec 06 07:18:44 crc kubenswrapper[4945]: I1206 07:18:44.295918 4945 generic.go:334] "Generic (PLEG): container finished" podID="72cf544a-1950-42b5-b244-b20339a8c702" containerID="724de5f0c3d669660e6698a3c99a4875f3d5e9150ef8166ff6741bfbba23a072" exitCode=0 Dec 06 07:18:44 crc kubenswrapper[4945]: I1206 07:18:44.296146 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zljkx" event={"ID":"72cf544a-1950-42b5-b244-b20339a8c702","Type":"ContainerDied","Data":"724de5f0c3d669660e6698a3c99a4875f3d5e9150ef8166ff6741bfbba23a072"} Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.629999 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.745584 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-combined-ca-bundle\") pod \"72cf544a-1950-42b5-b244-b20339a8c702\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.745635 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg8hp\" (UniqueName: \"kubernetes.io/projected/72cf544a-1950-42b5-b244-b20339a8c702-kube-api-access-gg8hp\") pod \"72cf544a-1950-42b5-b244-b20339a8c702\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.745998 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-config-data\") pod \"72cf544a-1950-42b5-b244-b20339a8c702\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.746079 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-scripts\") pod \"72cf544a-1950-42b5-b244-b20339a8c702\" (UID: \"72cf544a-1950-42b5-b244-b20339a8c702\") " Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.752272 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72cf544a-1950-42b5-b244-b20339a8c702-kube-api-access-gg8hp" (OuterVolumeSpecName: "kube-api-access-gg8hp") pod "72cf544a-1950-42b5-b244-b20339a8c702" (UID: "72cf544a-1950-42b5-b244-b20339a8c702"). InnerVolumeSpecName "kube-api-access-gg8hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.752858 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-scripts" (OuterVolumeSpecName: "scripts") pod "72cf544a-1950-42b5-b244-b20339a8c702" (UID: "72cf544a-1950-42b5-b244-b20339a8c702"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.778081 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-config-data" (OuterVolumeSpecName: "config-data") pod "72cf544a-1950-42b5-b244-b20339a8c702" (UID: "72cf544a-1950-42b5-b244-b20339a8c702"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.778148 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72cf544a-1950-42b5-b244-b20339a8c702" (UID: "72cf544a-1950-42b5-b244-b20339a8c702"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.848336 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.848392 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.848401 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72cf544a-1950-42b5-b244-b20339a8c702-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:45 crc kubenswrapper[4945]: I1206 07:18:45.848411 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg8hp\" (UniqueName: \"kubernetes.io/projected/72cf544a-1950-42b5-b244-b20339a8c702-kube-api-access-gg8hp\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.318900 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zljkx" event={"ID":"72cf544a-1950-42b5-b244-b20339a8c702","Type":"ContainerDied","Data":"eb292f3be12d664d670eea7c64176c074d16f5437ba2752cf434b73034d0fe6f"} Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.318951 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb292f3be12d664d670eea7c64176c074d16f5437ba2752cf434b73034d0fe6f" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.319458 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zljkx" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.369991 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.394763 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 07:18:46 crc kubenswrapper[4945]: E1206 07:18:46.395547 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72cf544a-1950-42b5-b244-b20339a8c702" containerName="nova-cell0-conductor-db-sync" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.395645 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72cf544a-1950-42b5-b244-b20339a8c702" containerName="nova-cell0-conductor-db-sync" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.395945 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72cf544a-1950-42b5-b244-b20339a8c702" containerName="nova-cell0-conductor-db-sync" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.396900 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.400312 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.400485 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-5tvtw" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.433012 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.563272 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcmkb\" (UniqueName: \"kubernetes.io/projected/72607de1-4f23-441e-b722-3844b2eb6d15-kube-api-access-mcmkb\") pod \"nova-cell0-conductor-0\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.563777 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.563846 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.666170 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.666338 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.666429 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcmkb\" (UniqueName: \"kubernetes.io/projected/72607de1-4f23-441e-b722-3844b2eb6d15-kube-api-access-mcmkb\") pod \"nova-cell0-conductor-0\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.671668 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.686815 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.687659 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcmkb\" (UniqueName: \"kubernetes.io/projected/72607de1-4f23-441e-b722-3844b2eb6d15-kube-api-access-mcmkb\") pod \"nova-cell0-conductor-0\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:46 crc kubenswrapper[4945]: I1206 07:18:46.715570 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:47 crc kubenswrapper[4945]: I1206 07:18:47.172430 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 07:18:47 crc kubenswrapper[4945]: I1206 07:18:47.332560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"72607de1-4f23-441e-b722-3844b2eb6d15","Type":"ContainerStarted","Data":"7e3e0f97c8029e1212380ce9fe49ecd37ce03778d08cf5c4792094828f4f3476"} Dec 06 07:18:48 crc kubenswrapper[4945]: I1206 07:18:48.347648 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"72607de1-4f23-441e-b722-3844b2eb6d15","Type":"ContainerStarted","Data":"fd592672f76730e847191d2934d9c9cfb67d21204f6a000e593a960e8df07a85"} Dec 06 07:18:48 crc kubenswrapper[4945]: I1206 07:18:48.348090 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:48 crc kubenswrapper[4945]: I1206 07:18:48.370917 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.370895752 podStartE2EDuration="2.370895752s" podCreationTimestamp="2025-12-06 07:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:18:48.364704625 +0000 UTC m=+1541.819565689" watchObservedRunningTime="2025-12-06 07:18:48.370895752 +0000 UTC m=+1541.825756796" Dec 06 07:18:50 crc kubenswrapper[4945]: I1206 07:18:50.293156 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:18:50 crc kubenswrapper[4945]: I1206 07:18:50.293792 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f5f1ef51-68e5-4133-a248-1d7251b5bb96" containerName="kube-state-metrics" containerID="cri-o://335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16" gracePeriod=30 Dec 06 07:18:50 crc kubenswrapper[4945]: I1206 07:18:50.849762 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.024445 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9w2f\" (UniqueName: \"kubernetes.io/projected/f5f1ef51-68e5-4133-a248-1d7251b5bb96-kube-api-access-z9w2f\") pod \"f5f1ef51-68e5-4133-a248-1d7251b5bb96\" (UID: \"f5f1ef51-68e5-4133-a248-1d7251b5bb96\") " Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.033341 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5f1ef51-68e5-4133-a248-1d7251b5bb96-kube-api-access-z9w2f" (OuterVolumeSpecName: "kube-api-access-z9w2f") pod "f5f1ef51-68e5-4133-a248-1d7251b5bb96" (UID: "f5f1ef51-68e5-4133-a248-1d7251b5bb96"). InnerVolumeSpecName "kube-api-access-z9w2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.128827 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9w2f\" (UniqueName: \"kubernetes.io/projected/f5f1ef51-68e5-4133-a248-1d7251b5bb96-kube-api-access-z9w2f\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.385599 4945 generic.go:334] "Generic (PLEG): container finished" podID="f5f1ef51-68e5-4133-a248-1d7251b5bb96" containerID="335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16" exitCode=2 Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.385662 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.385672 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f5f1ef51-68e5-4133-a248-1d7251b5bb96","Type":"ContainerDied","Data":"335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16"} Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.385756 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f5f1ef51-68e5-4133-a248-1d7251b5bb96","Type":"ContainerDied","Data":"8d0c5e2b19aa7a04906ff8be06ade8102a376be71b052d6e55158fe1764151ea"} Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.385805 4945 scope.go:117] "RemoveContainer" containerID="335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.427088 4945 scope.go:117] "RemoveContainer" containerID="335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16" Dec 06 07:18:51 crc kubenswrapper[4945]: E1206 07:18:51.428715 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16\": container with ID starting with 335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16 not found: ID does not exist" containerID="335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.428760 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16"} err="failed to get container status \"335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16\": rpc error: code = NotFound desc = could not find container \"335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16\": container with ID starting with 335be0bcfbd3df7e9d9b7597cfafb5e48985461774caa6852c699c3ca7b28b16 not found: ID does not exist" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.437912 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.454431 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.463352 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:18:51 crc kubenswrapper[4945]: E1206 07:18:51.463944 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5f1ef51-68e5-4133-a248-1d7251b5bb96" containerName="kube-state-metrics" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.463966 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5f1ef51-68e5-4133-a248-1d7251b5bb96" containerName="kube-state-metrics" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.464190 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5f1ef51-68e5-4133-a248-1d7251b5bb96" containerName="kube-state-metrics" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.464929 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.468021 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.468072 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.471971 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.537011 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.537087 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpzd7\" (UniqueName: \"kubernetes.io/projected/d5228457-326b-4470-ad26-f7bba1f50e18-kube-api-access-zpzd7\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.537293 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.537374 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.639328 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.639383 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpzd7\" (UniqueName: \"kubernetes.io/projected/d5228457-326b-4470-ad26-f7bba1f50e18-kube-api-access-zpzd7\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.640766 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.640897 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.645671 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.645784 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.647235 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.656271 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpzd7\" (UniqueName: \"kubernetes.io/projected/d5228457-326b-4470-ad26-f7bba1f50e18-kube-api-access-zpzd7\") pod \"kube-state-metrics-0\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " pod="openstack/kube-state-metrics-0" Dec 06 07:18:51 crc kubenswrapper[4945]: I1206 07:18:51.793452 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 07:18:52 crc kubenswrapper[4945]: W1206 07:18:52.314381 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5228457_326b_4470_ad26_f7bba1f50e18.slice/crio-4a66f7add2b33b5fbe9aeb2b42da28221ec0eda314e30a56ff1222f6c5b78ab3 WatchSource:0}: Error finding container 4a66f7add2b33b5fbe9aeb2b42da28221ec0eda314e30a56ff1222f6c5b78ab3: Status 404 returned error can't find the container with id 4a66f7add2b33b5fbe9aeb2b42da28221ec0eda314e30a56ff1222f6c5b78ab3 Dec 06 07:18:52 crc kubenswrapper[4945]: I1206 07:18:52.315208 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:18:52 crc kubenswrapper[4945]: I1206 07:18:52.404608 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d5228457-326b-4470-ad26-f7bba1f50e18","Type":"ContainerStarted","Data":"4a66f7add2b33b5fbe9aeb2b42da28221ec0eda314e30a56ff1222f6c5b78ab3"} Dec 06 07:18:52 crc kubenswrapper[4945]: I1206 07:18:52.428158 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:52 crc kubenswrapper[4945]: I1206 07:18:52.428491 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="ceilometer-central-agent" containerID="cri-o://cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1" gracePeriod=30 Dec 06 07:18:52 crc kubenswrapper[4945]: I1206 07:18:52.428553 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="proxy-httpd" containerID="cri-o://133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf" gracePeriod=30 Dec 06 07:18:52 crc kubenswrapper[4945]: I1206 07:18:52.428591 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="ceilometer-notification-agent" containerID="cri-o://92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065" gracePeriod=30 Dec 06 07:18:52 crc kubenswrapper[4945]: I1206 07:18:52.428554 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="sg-core" containerID="cri-o://b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6" gracePeriod=30 Dec 06 07:18:52 crc kubenswrapper[4945]: I1206 07:18:52.965741 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5f1ef51-68e5-4133-a248-1d7251b5bb96" path="/var/lib/kubelet/pods/f5f1ef51-68e5-4133-a248-1d7251b5bb96/volumes" Dec 06 07:18:53 crc kubenswrapper[4945]: I1206 07:18:53.416509 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d5228457-326b-4470-ad26-f7bba1f50e18","Type":"ContainerStarted","Data":"704c83d0821804990d78d0bdd43784abbc4a50001e79e73a668598b93f18f6ac"} Dec 06 07:18:53 crc kubenswrapper[4945]: I1206 07:18:53.416925 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 07:18:53 crc kubenswrapper[4945]: I1206 07:18:53.420317 4945 generic.go:334] "Generic (PLEG): container finished" podID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerID="133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf" exitCode=0 Dec 06 07:18:53 crc kubenswrapper[4945]: I1206 07:18:53.420348 4945 generic.go:334] "Generic (PLEG): container finished" podID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerID="b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6" exitCode=2 Dec 06 07:18:53 crc kubenswrapper[4945]: I1206 07:18:53.420357 4945 generic.go:334] "Generic (PLEG): container finished" podID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerID="cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1" exitCode=0 Dec 06 07:18:53 crc kubenswrapper[4945]: I1206 07:18:53.420377 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72d55594-ed65-4a69-b8cf-1a8925ea32a6","Type":"ContainerDied","Data":"133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf"} Dec 06 07:18:53 crc kubenswrapper[4945]: I1206 07:18:53.420416 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72d55594-ed65-4a69-b8cf-1a8925ea32a6","Type":"ContainerDied","Data":"b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6"} Dec 06 07:18:53 crc kubenswrapper[4945]: I1206 07:18:53.420428 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72d55594-ed65-4a69-b8cf-1a8925ea32a6","Type":"ContainerDied","Data":"cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1"} Dec 06 07:18:53 crc kubenswrapper[4945]: I1206 07:18:53.442665 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.006613185 podStartE2EDuration="2.442635492s" podCreationTimestamp="2025-12-06 07:18:51 +0000 UTC" firstStartedPulling="2025-12-06 07:18:52.318939572 +0000 UTC m=+1545.773800616" lastFinishedPulling="2025-12-06 07:18:52.754961889 +0000 UTC m=+1546.209822923" observedRunningTime="2025-12-06 07:18:53.434850531 +0000 UTC m=+1546.889711585" watchObservedRunningTime="2025-12-06 07:18:53.442635492 +0000 UTC m=+1546.897496536" Dec 06 07:18:56 crc kubenswrapper[4945]: I1206 07:18:56.786176 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.116050 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.285227 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsmvv\" (UniqueName: \"kubernetes.io/projected/72d55594-ed65-4a69-b8cf-1a8925ea32a6-kube-api-access-vsmvv\") pod \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.285604 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-run-httpd\") pod \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.285638 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-sg-core-conf-yaml\") pod \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.285700 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-config-data\") pod \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.286506 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-log-httpd\") pod \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.286584 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "72d55594-ed65-4a69-b8cf-1a8925ea32a6" (UID: "72d55594-ed65-4a69-b8cf-1a8925ea32a6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.286635 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-combined-ca-bundle\") pod \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.286669 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-scripts\") pod \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\" (UID: \"72d55594-ed65-4a69-b8cf-1a8925ea32a6\") " Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.286927 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "72d55594-ed65-4a69-b8cf-1a8925ea32a6" (UID: "72d55594-ed65-4a69-b8cf-1a8925ea32a6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.287410 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.287428 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72d55594-ed65-4a69-b8cf-1a8925ea32a6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.303932 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d55594-ed65-4a69-b8cf-1a8925ea32a6-kube-api-access-vsmvv" (OuterVolumeSpecName: "kube-api-access-vsmvv") pod "72d55594-ed65-4a69-b8cf-1a8925ea32a6" (UID: "72d55594-ed65-4a69-b8cf-1a8925ea32a6"). InnerVolumeSpecName "kube-api-access-vsmvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.308224 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-scripts" (OuterVolumeSpecName: "scripts") pod "72d55594-ed65-4a69-b8cf-1a8925ea32a6" (UID: "72d55594-ed65-4a69-b8cf-1a8925ea32a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.317204 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "72d55594-ed65-4a69-b8cf-1a8925ea32a6" (UID: "72d55594-ed65-4a69-b8cf-1a8925ea32a6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.366090 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72d55594-ed65-4a69-b8cf-1a8925ea32a6" (UID: "72d55594-ed65-4a69-b8cf-1a8925ea32a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.389913 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.390201 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.390319 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsmvv\" (UniqueName: \"kubernetes.io/projected/72d55594-ed65-4a69-b8cf-1a8925ea32a6-kube-api-access-vsmvv\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.390878 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.396383 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-config-data" (OuterVolumeSpecName: "config-data") pod "72d55594-ed65-4a69-b8cf-1a8925ea32a6" (UID: "72d55594-ed65-4a69-b8cf-1a8925ea32a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.480608 4945 generic.go:334] "Generic (PLEG): container finished" podID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerID="92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065" exitCode=0 Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.480667 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72d55594-ed65-4a69-b8cf-1a8925ea32a6","Type":"ContainerDied","Data":"92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065"} Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.480706 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72d55594-ed65-4a69-b8cf-1a8925ea32a6","Type":"ContainerDied","Data":"1fe9a5d504e096c96da60c21bcff141854a6ffeec4801aeef7693ee44aa93a75"} Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.480729 4945 scope.go:117] "RemoveContainer" containerID="133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.481171 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.531518 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-5j7ln"] Dec 06 07:18:57 crc kubenswrapper[4945]: E1206 07:18:57.532702 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="proxy-httpd" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.532828 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="proxy-httpd" Dec 06 07:18:57 crc kubenswrapper[4945]: E1206 07:18:57.532900 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="sg-core" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.532953 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="sg-core" Dec 06 07:18:57 crc kubenswrapper[4945]: E1206 07:18:57.533031 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="ceilometer-notification-agent" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.533084 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="ceilometer-notification-agent" Dec 06 07:18:57 crc kubenswrapper[4945]: E1206 07:18:57.533168 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="ceilometer-central-agent" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.533220 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="ceilometer-central-agent" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.533868 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="proxy-httpd" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.533987 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="ceilometer-central-agent" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.534076 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="ceilometer-notification-agent" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.534210 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" containerName="sg-core" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.535414 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.539885 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.540524 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.545615 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72d55594-ed65-4a69-b8cf-1a8925ea32a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.568403 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5j7ln"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.597462 4945 scope.go:117] "RemoveContainer" containerID="b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.597717 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.615732 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.647404 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-scripts\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.647512 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-config-data\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.647572 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkjdr\" (UniqueName: \"kubernetes.io/projected/b5afc31d-67d1-4faf-998c-97b660f3d510-kube-api-access-zkjdr\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.647638 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.668178 4945 scope.go:117] "RemoveContainer" containerID="92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.676206 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.679580 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.685802 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.686096 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.686302 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.689104 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.751177 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-scripts\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.751605 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78bmq\" (UniqueName: \"kubernetes.io/projected/b1eb7e6e-6348-4285-81ff-ae05307cc91e-kube-api-access-78bmq\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.751766 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-scripts\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.751990 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.752144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-config-data\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.752327 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.752466 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-log-httpd\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.752613 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkjdr\" (UniqueName: \"kubernetes.io/projected/b5afc31d-67d1-4faf-998c-97b660f3d510-kube-api-access-zkjdr\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.753036 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-config-data\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.753818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.754087 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.759897 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-scripts\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.760349 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-run-httpd\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.768007 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-config-data\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.770033 4945 scope.go:117] "RemoveContainer" containerID="cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.771056 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.778915 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.781815 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.791105 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.802922 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.819674 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkjdr\" (UniqueName: \"kubernetes.io/projected/b5afc31d-67d1-4faf-998c-97b660f3d510-kube-api-access-zkjdr\") pod \"nova-cell0-cell-mapping-5j7ln\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.838343 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.854195 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.858068 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.862676 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-run-httpd\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.862741 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78bmq\" (UniqueName: \"kubernetes.io/projected/b1eb7e6e-6348-4285-81ff-ae05307cc91e-kube-api-access-78bmq\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.862788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-scripts\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.862809 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-config-data\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.862860 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.862902 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.862921 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-log-httpd\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.862956 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-config-data\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.862982 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aa39339-6d1c-43db-87dd-ae3da33013f1-logs\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.863009 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.863069 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.863130 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szb9b\" (UniqueName: \"kubernetes.io/projected/0aa39339-6d1c-43db-87dd-ae3da33013f1-kube-api-access-szb9b\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.863738 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-run-httpd\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.871257 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-log-httpd\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.871569 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.881153 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-scripts\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.881639 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.884967 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.888200 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.889059 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.901018 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78bmq\" (UniqueName: \"kubernetes.io/projected/b1eb7e6e-6348-4285-81ff-ae05307cc91e-kube-api-access-78bmq\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.906367 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.908208 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.913060 4945 scope.go:117] "RemoveContainer" containerID="133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.913328 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-config-data\") pod \"ceilometer-0\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " pod="openstack/ceilometer-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.913752 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 07:18:57 crc kubenswrapper[4945]: E1206 07:18:57.934139 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf\": container with ID starting with 133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf not found: ID does not exist" containerID="133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.934541 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf"} err="failed to get container status \"133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf\": rpc error: code = NotFound desc = could not find container \"133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf\": container with ID starting with 133e408c6f4606985ba5da4a964ec7e92b46d260c393981e26d0b523ea94baaf not found: ID does not exist" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.934667 4945 scope.go:117] "RemoveContainer" containerID="b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6" Dec 06 07:18:57 crc kubenswrapper[4945]: E1206 07:18:57.941482 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6\": container with ID starting with b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6 not found: ID does not exist" containerID="b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.941554 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6"} err="failed to get container status \"b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6\": rpc error: code = NotFound desc = could not find container \"b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6\": container with ID starting with b6602905241628c59c237b591e4eb9433393d7890fb7041f1bd394e736c811b6 not found: ID does not exist" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.941590 4945 scope.go:117] "RemoveContainer" containerID="92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065" Dec 06 07:18:57 crc kubenswrapper[4945]: E1206 07:18:57.942232 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065\": container with ID starting with 92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065 not found: ID does not exist" containerID="92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.942330 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065"} err="failed to get container status \"92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065\": rpc error: code = NotFound desc = could not find container \"92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065\": container with ID starting with 92585ae415c1e774887e6075f4113700bce9fe1a247de609a58a9fa0c5ba8065 not found: ID does not exist" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.942387 4945 scope.go:117] "RemoveContainer" containerID="cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1" Dec 06 07:18:57 crc kubenswrapper[4945]: E1206 07:18:57.942728 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1\": container with ID starting with cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1 not found: ID does not exist" containerID="cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.942752 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1"} err="failed to get container status \"cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1\": rpc error: code = NotFound desc = could not find container \"cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1\": container with ID starting with cf698010cfd3515b7a4bebfcb85d99139da2648a8ea9b88849b820f339fb8ea1 not found: ID does not exist" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.970923 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.972144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aa39339-6d1c-43db-87dd-ae3da33013f1-logs\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.972193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.972254 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffxkj\" (UniqueName: \"kubernetes.io/projected/54d6c3b2-3db2-4400-958e-839a0f53cda0-kube-api-access-ffxkj\") pod \"nova-cell1-novncproxy-0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.972302 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.972322 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffq46\" (UniqueName: \"kubernetes.io/projected/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-kube-api-access-ffq46\") pod \"nova-scheduler-0\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " pod="openstack/nova-scheduler-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.972358 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szb9b\" (UniqueName: \"kubernetes.io/projected/0aa39339-6d1c-43db-87dd-ae3da33013f1-kube-api-access-szb9b\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.972384 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-config-data\") pod \"nova-scheduler-0\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " pod="openstack/nova-scheduler-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.972419 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " pod="openstack/nova-scheduler-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.972451 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-config-data\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.972477 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.973010 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aa39339-6d1c-43db-87dd-ae3da33013f1-logs\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.982308 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-config-data\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:57 crc kubenswrapper[4945]: I1206 07:18:57.982739 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.004537 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szb9b\" (UniqueName: \"kubernetes.io/projected/0aa39339-6d1c-43db-87dd-ae3da33013f1-kube-api-access-szb9b\") pod \"nova-api-0\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " pod="openstack/nova-api-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.008716 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.046185 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.055400 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.067538 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.074043 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-config-data\") pod \"nova-scheduler-0\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " pod="openstack/nova-scheduler-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.075645 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " pod="openstack/nova-scheduler-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.075696 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.075864 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffxkj\" (UniqueName: \"kubernetes.io/projected/54d6c3b2-3db2-4400-958e-839a0f53cda0-kube-api-access-ffxkj\") pod \"nova-cell1-novncproxy-0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.075900 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.075920 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffq46\" (UniqueName: \"kubernetes.io/projected/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-kube-api-access-ffq46\") pod \"nova-scheduler-0\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " pod="openstack/nova-scheduler-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.077565 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.079862 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-config-data\") pod \"nova-scheduler-0\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " pod="openstack/nova-scheduler-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.122017 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.131990 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffq46\" (UniqueName: \"kubernetes.io/projected/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-kube-api-access-ffq46\") pod \"nova-scheduler-0\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " pod="openstack/nova-scheduler-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.135885 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.138823 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffxkj\" (UniqueName: \"kubernetes.io/projected/54d6c3b2-3db2-4400-958e-839a0f53cda0-kube-api-access-ffxkj\") pod \"nova-cell1-novncproxy-0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.146740 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " pod="openstack/nova-scheduler-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.147342 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.182239 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d313433-9f21-44c9-94af-49d0b419af57-logs\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.182333 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.182389 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cthnv\" (UniqueName: \"kubernetes.io/projected/2d313433-9f21-44c9-94af-49d0b419af57-kube-api-access-cthnv\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.182420 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-config-data\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.225393 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-fhcms"] Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.235594 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.256878 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-fhcms"] Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.298557 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d313433-9f21-44c9-94af-49d0b419af57-logs\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.298611 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-nb\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.298675 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-config\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.298733 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.298754 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-sb\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.298849 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cthnv\" (UniqueName: \"kubernetes.io/projected/2d313433-9f21-44c9-94af-49d0b419af57-kube-api-access-cthnv\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.298876 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksmqf\" (UniqueName: \"kubernetes.io/projected/445042de-1881-4990-bf51-d77ec753f986-kube-api-access-ksmqf\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.298897 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-config-data\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.299033 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-svc\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.299076 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-swift-storage-0\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.311518 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d313433-9f21-44c9-94af-49d0b419af57-logs\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.324818 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.325224 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-config-data\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.333111 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cthnv\" (UniqueName: \"kubernetes.io/projected/2d313433-9f21-44c9-94af-49d0b419af57-kube-api-access-cthnv\") pod \"nova-metadata-0\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.352627 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.374417 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.414957 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksmqf\" (UniqueName: \"kubernetes.io/projected/445042de-1881-4990-bf51-d77ec753f986-kube-api-access-ksmqf\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.415052 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-svc\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.415083 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-swift-storage-0\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.415178 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-nb\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.415206 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-config\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.415236 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-sb\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.416377 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-sb\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.418134 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-svc\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.419414 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-nb\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.421233 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-config\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.421629 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-swift-storage-0\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.429087 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.480250 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksmqf\" (UniqueName: \"kubernetes.io/projected/445042de-1881-4990-bf51-d77ec753f986-kube-api-access-ksmqf\") pod \"dnsmasq-dns-7d75688ddc-fhcms\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.599363 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.885705 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5j7ln"] Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.926443 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:18:58 crc kubenswrapper[4945]: I1206 07:18:58.992148 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72d55594-ed65-4a69-b8cf-1a8925ea32a6" path="/var/lib/kubelet/pods/72d55594-ed65-4a69-b8cf-1a8925ea32a6/volumes" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.044836 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kd5fs"] Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.053703 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.057997 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.060831 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.075698 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kd5fs"] Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.135824 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxqbf\" (UniqueName: \"kubernetes.io/projected/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-kube-api-access-zxqbf\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.135921 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-scripts\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.135970 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.136082 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-config-data\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.166398 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.185392 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.227742 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.239042 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-scripts\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.239120 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.239217 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-config-data\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.239313 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxqbf\" (UniqueName: \"kubernetes.io/projected/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-kube-api-access-zxqbf\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.250828 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.251981 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-config-data\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.255412 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-scripts\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.260273 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxqbf\" (UniqueName: \"kubernetes.io/projected/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-kube-api-access-zxqbf\") pod \"nova-cell1-conductor-db-sync-kd5fs\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.387765 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-fhcms"] Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.400164 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:18:59 crc kubenswrapper[4945]: W1206 07:18:59.409578 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54d6c3b2_3db2_4400_958e_839a0f53cda0.slice/crio-0ed337f4ef64726823d490c493b0dd3c7f12e9aa048c8332fe01a3bf01c73a1e WatchSource:0}: Error finding container 0ed337f4ef64726823d490c493b0dd3c7f12e9aa048c8332fe01a3bf01c73a1e: Status 404 returned error can't find the container with id 0ed337f4ef64726823d490c493b0dd3c7f12e9aa048c8332fe01a3bf01c73a1e Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.430519 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.564907 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a3edb3c2-9c23-47e2-a243-dc871eef8e8c","Type":"ContainerStarted","Data":"23f1fe5170291dd699701a9d8a8b954d4d5504453e28500637c032647607a23c"} Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.566602 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0aa39339-6d1c-43db-87dd-ae3da33013f1","Type":"ContainerStarted","Data":"205eeeedd1b5da38fa36f0f645db3488e1bd859608ef97d33991a4c6dd8b0abf"} Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.568269 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5j7ln" event={"ID":"b5afc31d-67d1-4faf-998c-97b660f3d510","Type":"ContainerStarted","Data":"2f935bfecfa8a7af798c6c8d352ae3b0ed1c935c6dba724d358a31059be6f477"} Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.568323 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5j7ln" event={"ID":"b5afc31d-67d1-4faf-998c-97b660f3d510","Type":"ContainerStarted","Data":"5714e8c5e27f885b24bab90dac85b8c7d0c399c4d5cf6c32a03479dc8e300061"} Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.570644 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1eb7e6e-6348-4285-81ff-ae05307cc91e","Type":"ContainerStarted","Data":"7389dc1f6c51556cd21183504f03406664433f35936be7115037f2b47bade53b"} Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.587368 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"54d6c3b2-3db2-4400-958e-839a0f53cda0","Type":"ContainerStarted","Data":"0ed337f4ef64726823d490c493b0dd3c7f12e9aa048c8332fe01a3bf01c73a1e"} Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.599885 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" event={"ID":"445042de-1881-4990-bf51-d77ec753f986","Type":"ContainerStarted","Data":"52eea5c2e18b5c0f7b7c7fa217b419de8092de32c9ebb410cf7e11e6ecd68ee8"} Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.605533 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2d313433-9f21-44c9-94af-49d0b419af57","Type":"ContainerStarted","Data":"4850c16ab7cadc8b460be830692cf4b0089f61d862c6c7ed73e153b71a0307ae"} Dec 06 07:18:59 crc kubenswrapper[4945]: I1206 07:18:59.619919 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-5j7ln" podStartSLOduration=2.619890742 podStartE2EDuration="2.619890742s" podCreationTimestamp="2025-12-06 07:18:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:18:59.602140662 +0000 UTC m=+1553.057001716" watchObservedRunningTime="2025-12-06 07:18:59.619890742 +0000 UTC m=+1553.074751786" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.025778 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kd5fs"] Dec 06 07:19:00 crc kubenswrapper[4945]: W1206 07:19:00.034202 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8b7d3fd_fcf0_4f38_b2d1_31b883b0922b.slice/crio-e2789f73a3945a75266c91c64626f326df5fbcf70337bc683825b5f7a932c687 WatchSource:0}: Error finding container e2789f73a3945a75266c91c64626f326df5fbcf70337bc683825b5f7a932c687: Status 404 returned error can't find the container with id e2789f73a3945a75266c91c64626f326df5fbcf70337bc683825b5f7a932c687 Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.689350 4945 generic.go:334] "Generic (PLEG): container finished" podID="445042de-1881-4990-bf51-d77ec753f986" containerID="73f7cd0ccf609d91c0734d2092d72f5d6c6161a5f7adc0ef19b7dc82a92539f4" exitCode=0 Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.689781 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" event={"ID":"445042de-1881-4990-bf51-d77ec753f986","Type":"ContainerDied","Data":"73f7cd0ccf609d91c0734d2092d72f5d6c6161a5f7adc0ef19b7dc82a92539f4"} Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.691831 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vxm6x"] Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.695187 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.706798 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kd5fs" event={"ID":"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b","Type":"ContainerStarted","Data":"0f377c0a02117dc7052bb48a4af37d61f3095dd82d7bde891095afbd858d61ee"} Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.706856 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kd5fs" event={"ID":"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b","Type":"ContainerStarted","Data":"e2789f73a3945a75266c91c64626f326df5fbcf70337bc683825b5f7a932c687"} Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.720171 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1eb7e6e-6348-4285-81ff-ae05307cc91e","Type":"ContainerStarted","Data":"b9af4bd5a62cda53dd725a20e1df6be731fdc58e780f4838726be7126c072f5a"} Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.788359 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vxm6x"] Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.831129 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-utilities\") pod \"community-operators-vxm6x\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.831188 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-catalog-content\") pod \"community-operators-vxm6x\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.831311 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vb4l\" (UniqueName: \"kubernetes.io/projected/113b0590-9eee-4220-b7c9-44fcb69f250f-kube-api-access-7vb4l\") pod \"community-operators-vxm6x\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.875686 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-kd5fs" podStartSLOduration=2.8756554850000002 podStartE2EDuration="2.875655485s" podCreationTimestamp="2025-12-06 07:18:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:00.777704104 +0000 UTC m=+1554.232565179" watchObservedRunningTime="2025-12-06 07:19:00.875655485 +0000 UTC m=+1554.330516529" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.933312 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-utilities\") pod \"community-operators-vxm6x\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.933370 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-catalog-content\") pod \"community-operators-vxm6x\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.933433 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vb4l\" (UniqueName: \"kubernetes.io/projected/113b0590-9eee-4220-b7c9-44fcb69f250f-kube-api-access-7vb4l\") pod \"community-operators-vxm6x\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.933966 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-utilities\") pod \"community-operators-vxm6x\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.934135 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-catalog-content\") pod \"community-operators-vxm6x\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:00 crc kubenswrapper[4945]: I1206 07:19:00.955316 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vb4l\" (UniqueName: \"kubernetes.io/projected/113b0590-9eee-4220-b7c9-44fcb69f250f-kube-api-access-7vb4l\") pod \"community-operators-vxm6x\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:01 crc kubenswrapper[4945]: I1206 07:19:01.070232 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:01 crc kubenswrapper[4945]: I1206 07:19:01.812433 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 07:19:03 crc kubenswrapper[4945]: I1206 07:19:03.705665 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:03 crc kubenswrapper[4945]: I1206 07:19:03.781331 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:19:04 crc kubenswrapper[4945]: I1206 07:19:04.783571 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vxm6x"] Dec 06 07:19:04 crc kubenswrapper[4945]: W1206 07:19:04.787373 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod113b0590_9eee_4220_b7c9_44fcb69f250f.slice/crio-55f4d5300e414665103682964b766dae6291dd9175b1e421d8e55f3880d04f22 WatchSource:0}: Error finding container 55f4d5300e414665103682964b766dae6291dd9175b1e421d8e55f3880d04f22: Status 404 returned error can't find the container with id 55f4d5300e414665103682964b766dae6291dd9175b1e421d8e55f3880d04f22 Dec 06 07:19:04 crc kubenswrapper[4945]: I1206 07:19:04.792066 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a3edb3c2-9c23-47e2-a243-dc871eef8e8c","Type":"ContainerStarted","Data":"07ed647e4210a85928967161b01d1a619e42389af3b1ff7678ac084276443c77"} Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.815578 4945 generic.go:334] "Generic (PLEG): container finished" podID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerID="40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1" exitCode=0 Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.816233 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vxm6x" event={"ID":"113b0590-9eee-4220-b7c9-44fcb69f250f","Type":"ContainerDied","Data":"40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1"} Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.816615 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vxm6x" event={"ID":"113b0590-9eee-4220-b7c9-44fcb69f250f","Type":"ContainerStarted","Data":"55f4d5300e414665103682964b766dae6291dd9175b1e421d8e55f3880d04f22"} Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.829254 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1eb7e6e-6348-4285-81ff-ae05307cc91e","Type":"ContainerStarted","Data":"5ed82833aca583fbab32a527da4bef5b512b77c7c64397f7742dbe74a40921e4"} Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.848983 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"54d6c3b2-3db2-4400-958e-839a0f53cda0","Type":"ContainerStarted","Data":"0ab1417aa10f14854712f6287bc7ff98a77fec3e4030e9d198a868daeb2bd9f6"} Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.851128 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="54d6c3b2-3db2-4400-958e-839a0f53cda0" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://0ab1417aa10f14854712f6287bc7ff98a77fec3e4030e9d198a868daeb2bd9f6" gracePeriod=30 Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.858703 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.796146225 podStartE2EDuration="8.858669385s" podCreationTimestamp="2025-12-06 07:18:57 +0000 UTC" firstStartedPulling="2025-12-06 07:18:59.184719979 +0000 UTC m=+1552.639581023" lastFinishedPulling="2025-12-06 07:19:04.247243139 +0000 UTC m=+1557.702104183" observedRunningTime="2025-12-06 07:19:04.82314029 +0000 UTC m=+1558.278001334" watchObservedRunningTime="2025-12-06 07:19:05.858669385 +0000 UTC m=+1559.313530429" Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.864584 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" event={"ID":"445042de-1881-4990-bf51-d77ec753f986","Type":"ContainerStarted","Data":"8dd28a205c7b31da8ae18ba488911ee5eb0dc3e4a56c2a8ca3c0445a6e99a115"} Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.865724 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.869528 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2d313433-9f21-44c9-94af-49d0b419af57","Type":"ContainerStarted","Data":"9ad74fd0080d46ac75c8e9c4d5e1afbe12423d9b9faf094925f976e70a1c11e0"} Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.869558 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2d313433-9f21-44c9-94af-49d0b419af57","Type":"ContainerStarted","Data":"577f670f82aa3dce9242bce18180272ae3d7e8c153cb06c9364212f739826691"} Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.869684 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2d313433-9f21-44c9-94af-49d0b419af57" containerName="nova-metadata-log" containerID="cri-o://577f670f82aa3dce9242bce18180272ae3d7e8c153cb06c9364212f739826691" gracePeriod=30 Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.869786 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2d313433-9f21-44c9-94af-49d0b419af57" containerName="nova-metadata-metadata" containerID="cri-o://9ad74fd0080d46ac75c8e9c4d5e1afbe12423d9b9faf094925f976e70a1c11e0" gracePeriod=30 Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.881513 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0aa39339-6d1c-43db-87dd-ae3da33013f1","Type":"ContainerStarted","Data":"087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718"} Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.881589 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0aa39339-6d1c-43db-87dd-ae3da33013f1","Type":"ContainerStarted","Data":"a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4"} Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.897354 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=4.062217623 podStartE2EDuration="8.897329371s" podCreationTimestamp="2025-12-06 07:18:57 +0000 UTC" firstStartedPulling="2025-12-06 07:18:59.418542915 +0000 UTC m=+1552.873403959" lastFinishedPulling="2025-12-06 07:19:04.253654673 +0000 UTC m=+1557.708515707" observedRunningTime="2025-12-06 07:19:05.87401343 +0000 UTC m=+1559.328874474" watchObservedRunningTime="2025-12-06 07:19:05.897329371 +0000 UTC m=+1559.352190415" Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.914478 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.880462026 podStartE2EDuration="7.914458685s" podCreationTimestamp="2025-12-06 07:18:58 +0000 UTC" firstStartedPulling="2025-12-06 07:18:59.23763636 +0000 UTC m=+1552.692497404" lastFinishedPulling="2025-12-06 07:19:04.271633019 +0000 UTC m=+1557.726494063" observedRunningTime="2025-12-06 07:19:05.907493996 +0000 UTC m=+1559.362355060" watchObservedRunningTime="2025-12-06 07:19:05.914458685 +0000 UTC m=+1559.369319719" Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.941192 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" podStartSLOduration=7.941170317 podStartE2EDuration="7.941170317s" podCreationTimestamp="2025-12-06 07:18:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:05.928602867 +0000 UTC m=+1559.383463911" watchObservedRunningTime="2025-12-06 07:19:05.941170317 +0000 UTC m=+1559.396031361" Dec 06 07:19:05 crc kubenswrapper[4945]: I1206 07:19:05.956408 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.61377054 podStartE2EDuration="8.956383189s" podCreationTimestamp="2025-12-06 07:18:57 +0000 UTC" firstStartedPulling="2025-12-06 07:18:58.92900113 +0000 UTC m=+1552.383862174" lastFinishedPulling="2025-12-06 07:19:04.271613779 +0000 UTC m=+1557.726474823" observedRunningTime="2025-12-06 07:19:05.949086481 +0000 UTC m=+1559.403947525" watchObservedRunningTime="2025-12-06 07:19:05.956383189 +0000 UTC m=+1559.411244223" Dec 06 07:19:06 crc kubenswrapper[4945]: I1206 07:19:06.897193 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1eb7e6e-6348-4285-81ff-ae05307cc91e","Type":"ContainerStarted","Data":"c36d1498eada2e39ab90b6f41041b808f19d09761fcb2090a1abd248c195a6c3"} Dec 06 07:19:06 crc kubenswrapper[4945]: I1206 07:19:06.900742 4945 generic.go:334] "Generic (PLEG): container finished" podID="2d313433-9f21-44c9-94af-49d0b419af57" containerID="9ad74fd0080d46ac75c8e9c4d5e1afbe12423d9b9faf094925f976e70a1c11e0" exitCode=0 Dec 06 07:19:06 crc kubenswrapper[4945]: I1206 07:19:06.900788 4945 generic.go:334] "Generic (PLEG): container finished" podID="2d313433-9f21-44c9-94af-49d0b419af57" containerID="577f670f82aa3dce9242bce18180272ae3d7e8c153cb06c9364212f739826691" exitCode=143 Dec 06 07:19:06 crc kubenswrapper[4945]: I1206 07:19:06.900850 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2d313433-9f21-44c9-94af-49d0b419af57","Type":"ContainerDied","Data":"9ad74fd0080d46ac75c8e9c4d5e1afbe12423d9b9faf094925f976e70a1c11e0"} Dec 06 07:19:06 crc kubenswrapper[4945]: I1206 07:19:06.900889 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2d313433-9f21-44c9-94af-49d0b419af57","Type":"ContainerDied","Data":"577f670f82aa3dce9242bce18180272ae3d7e8c153cb06c9364212f739826691"} Dec 06 07:19:06 crc kubenswrapper[4945]: I1206 07:19:06.910849 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vxm6x" event={"ID":"113b0590-9eee-4220-b7c9-44fcb69f250f","Type":"ContainerStarted","Data":"8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74"} Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.358343 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.559493 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-config-data\") pod \"2d313433-9f21-44c9-94af-49d0b419af57\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.559684 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cthnv\" (UniqueName: \"kubernetes.io/projected/2d313433-9f21-44c9-94af-49d0b419af57-kube-api-access-cthnv\") pod \"2d313433-9f21-44c9-94af-49d0b419af57\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.559737 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-combined-ca-bundle\") pod \"2d313433-9f21-44c9-94af-49d0b419af57\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.559766 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d313433-9f21-44c9-94af-49d0b419af57-logs\") pod \"2d313433-9f21-44c9-94af-49d0b419af57\" (UID: \"2d313433-9f21-44c9-94af-49d0b419af57\") " Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.560191 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d313433-9f21-44c9-94af-49d0b419af57-logs" (OuterVolumeSpecName: "logs") pod "2d313433-9f21-44c9-94af-49d0b419af57" (UID: "2d313433-9f21-44c9-94af-49d0b419af57"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.560477 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d313433-9f21-44c9-94af-49d0b419af57-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.564475 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d313433-9f21-44c9-94af-49d0b419af57-kube-api-access-cthnv" (OuterVolumeSpecName: "kube-api-access-cthnv") pod "2d313433-9f21-44c9-94af-49d0b419af57" (UID: "2d313433-9f21-44c9-94af-49d0b419af57"). InnerVolumeSpecName "kube-api-access-cthnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.590799 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d313433-9f21-44c9-94af-49d0b419af57" (UID: "2d313433-9f21-44c9-94af-49d0b419af57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.594590 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-config-data" (OuterVolumeSpecName: "config-data") pod "2d313433-9f21-44c9-94af-49d0b419af57" (UID: "2d313433-9f21-44c9-94af-49d0b419af57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.662760 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cthnv\" (UniqueName: \"kubernetes.io/projected/2d313433-9f21-44c9-94af-49d0b419af57-kube-api-access-cthnv\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.662801 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.662811 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d313433-9f21-44c9-94af-49d0b419af57-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.931431 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2d313433-9f21-44c9-94af-49d0b419af57","Type":"ContainerDied","Data":"4850c16ab7cadc8b460be830692cf4b0089f61d862c6c7ed73e153b71a0307ae"} Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.931498 4945 scope.go:117] "RemoveContainer" containerID="9ad74fd0080d46ac75c8e9c4d5e1afbe12423d9b9faf094925f976e70a1c11e0" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.931639 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.946698 4945 generic.go:334] "Generic (PLEG): container finished" podID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerID="8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74" exitCode=0 Dec 06 07:19:07 crc kubenswrapper[4945]: I1206 07:19:07.948107 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vxm6x" event={"ID":"113b0590-9eee-4220-b7c9-44fcb69f250f","Type":"ContainerDied","Data":"8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74"} Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.002478 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.008762 4945 scope.go:117] "RemoveContainer" containerID="577f670f82aa3dce9242bce18180272ae3d7e8c153cb06c9364212f739826691" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.014898 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.014946 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.034300 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.051333 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:08 crc kubenswrapper[4945]: E1206 07:19:08.051947 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d313433-9f21-44c9-94af-49d0b419af57" containerName="nova-metadata-log" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.052003 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d313433-9f21-44c9-94af-49d0b419af57" containerName="nova-metadata-log" Dec 06 07:19:08 crc kubenswrapper[4945]: E1206 07:19:08.052070 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d313433-9f21-44c9-94af-49d0b419af57" containerName="nova-metadata-metadata" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.052089 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d313433-9f21-44c9-94af-49d0b419af57" containerName="nova-metadata-metadata" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.052371 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d313433-9f21-44c9-94af-49d0b419af57" containerName="nova-metadata-metadata" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.052406 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d313433-9f21-44c9-94af-49d0b419af57" containerName="nova-metadata-log" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.054585 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.059495 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.059881 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.099164 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-config-data\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.099255 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a61e6da-dded-4d12-b4f2-1907f0647817-logs\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.099423 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwmzb\" (UniqueName: \"kubernetes.io/projected/6a61e6da-dded-4d12-b4f2-1907f0647817-kube-api-access-rwmzb\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.099516 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.099659 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.101334 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.202533 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a61e6da-dded-4d12-b4f2-1907f0647817-logs\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.202657 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwmzb\" (UniqueName: \"kubernetes.io/projected/6a61e6da-dded-4d12-b4f2-1907f0647817-kube-api-access-rwmzb\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.203040 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a61e6da-dded-4d12-b4f2-1907f0647817-logs\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.203392 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.203592 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.203747 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-config-data\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.210120 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.210148 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.220119 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-config-data\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.225007 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwmzb\" (UniqueName: \"kubernetes.io/projected/6a61e6da-dded-4d12-b4f2-1907f0647817-kube-api-access-rwmzb\") pod \"nova-metadata-0\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.353753 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.353796 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.375790 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.388124 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.388240 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.992301 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d313433-9f21-44c9-94af-49d0b419af57" path="/var/lib/kubelet/pods/2d313433-9f21-44c9-94af-49d0b419af57/volumes" Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.995170 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1eb7e6e-6348-4285-81ff-ae05307cc91e","Type":"ContainerStarted","Data":"58a37447d6a7e82e60ac8bec54e4492001005996d111a5d73ac2f6fa2b4804b7"} Dec 06 07:19:08 crc kubenswrapper[4945]: I1206 07:19:08.995226 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.010449 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.028385 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.169541935 podStartE2EDuration="12.028351627s" podCreationTimestamp="2025-12-06 07:18:57 +0000 UTC" firstStartedPulling="2025-12-06 07:18:59.230428295 +0000 UTC m=+1552.685289339" lastFinishedPulling="2025-12-06 07:19:07.089237987 +0000 UTC m=+1560.544099031" observedRunningTime="2025-12-06 07:19:09.012868139 +0000 UTC m=+1562.467729183" watchObservedRunningTime="2025-12-06 07:19:09.028351627 +0000 UTC m=+1562.483212671" Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.040887 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.101678 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.102582 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.978625 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vxm6x" event={"ID":"113b0590-9eee-4220-b7c9-44fcb69f250f","Type":"ContainerStarted","Data":"07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa"} Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.983722 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a61e6da-dded-4d12-b4f2-1907f0647817","Type":"ContainerStarted","Data":"ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170"} Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.983787 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a61e6da-dded-4d12-b4f2-1907f0647817","Type":"ContainerStarted","Data":"3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df"} Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.983804 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a61e6da-dded-4d12-b4f2-1907f0647817","Type":"ContainerStarted","Data":"c4660b40579cb0907fa70d1e8612c3403244e3c8bd277554990c3bd23e68eaf0"} Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.988213 4945 generic.go:334] "Generic (PLEG): container finished" podID="b5afc31d-67d1-4faf-998c-97b660f3d510" containerID="2f935bfecfa8a7af798c6c8d352ae3b0ed1c935c6dba724d358a31059be6f477" exitCode=0 Dec 06 07:19:09 crc kubenswrapper[4945]: I1206 07:19:09.988423 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5j7ln" event={"ID":"b5afc31d-67d1-4faf-998c-97b660f3d510","Type":"ContainerDied","Data":"2f935bfecfa8a7af798c6c8d352ae3b0ed1c935c6dba724d358a31059be6f477"} Dec 06 07:19:10 crc kubenswrapper[4945]: I1206 07:19:10.018864 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vxm6x" podStartSLOduration=6.857189409 podStartE2EDuration="10.018831624s" podCreationTimestamp="2025-12-06 07:19:00 +0000 UTC" firstStartedPulling="2025-12-06 07:19:05.837384269 +0000 UTC m=+1559.292245313" lastFinishedPulling="2025-12-06 07:19:08.999026484 +0000 UTC m=+1562.453887528" observedRunningTime="2025-12-06 07:19:10.004744943 +0000 UTC m=+1563.459605997" watchObservedRunningTime="2025-12-06 07:19:10.018831624 +0000 UTC m=+1563.473692668" Dec 06 07:19:10 crc kubenswrapper[4945]: I1206 07:19:10.102457 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.102427206 podStartE2EDuration="3.102427206s" podCreationTimestamp="2025-12-06 07:19:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:10.062540767 +0000 UTC m=+1563.517401831" watchObservedRunningTime="2025-12-06 07:19:10.102427206 +0000 UTC m=+1563.557288250" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.070856 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.070923 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.408420 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.499472 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-combined-ca-bundle\") pod \"b5afc31d-67d1-4faf-998c-97b660f3d510\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.499561 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkjdr\" (UniqueName: \"kubernetes.io/projected/b5afc31d-67d1-4faf-998c-97b660f3d510-kube-api-access-zkjdr\") pod \"b5afc31d-67d1-4faf-998c-97b660f3d510\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.499604 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-scripts\") pod \"b5afc31d-67d1-4faf-998c-97b660f3d510\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.499782 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-config-data\") pod \"b5afc31d-67d1-4faf-998c-97b660f3d510\" (UID: \"b5afc31d-67d1-4faf-998c-97b660f3d510\") " Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.507231 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5afc31d-67d1-4faf-998c-97b660f3d510-kube-api-access-zkjdr" (OuterVolumeSpecName: "kube-api-access-zkjdr") pod "b5afc31d-67d1-4faf-998c-97b660f3d510" (UID: "b5afc31d-67d1-4faf-998c-97b660f3d510"). InnerVolumeSpecName "kube-api-access-zkjdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.508000 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-scripts" (OuterVolumeSpecName: "scripts") pod "b5afc31d-67d1-4faf-998c-97b660f3d510" (UID: "b5afc31d-67d1-4faf-998c-97b660f3d510"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.534644 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-config-data" (OuterVolumeSpecName: "config-data") pod "b5afc31d-67d1-4faf-998c-97b660f3d510" (UID: "b5afc31d-67d1-4faf-998c-97b660f3d510"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.547246 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5afc31d-67d1-4faf-998c-97b660f3d510" (UID: "b5afc31d-67d1-4faf-998c-97b660f3d510"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.602348 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.602388 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.602402 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkjdr\" (UniqueName: \"kubernetes.io/projected/b5afc31d-67d1-4faf-998c-97b660f3d510-kube-api-access-zkjdr\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:11 crc kubenswrapper[4945]: I1206 07:19:11.602414 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5afc31d-67d1-4faf-998c-97b660f3d510-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.008059 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5j7ln" Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.008059 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5j7ln" event={"ID":"b5afc31d-67d1-4faf-998c-97b660f3d510","Type":"ContainerDied","Data":"5714e8c5e27f885b24bab90dac85b8c7d0c399c4d5cf6c32a03479dc8e300061"} Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.008115 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5714e8c5e27f885b24bab90dac85b8c7d0c399c4d5cf6c32a03479dc8e300061" Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.131086 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-vxm6x" podUID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerName="registry-server" probeResult="failure" output=< Dec 06 07:19:12 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 07:19:12 crc kubenswrapper[4945]: > Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.279167 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.279602 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerName="nova-api-log" containerID="cri-o://a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4" gracePeriod=30 Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.279799 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerName="nova-api-api" containerID="cri-o://087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718" gracePeriod=30 Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.296585 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.296848 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a3edb3c2-9c23-47e2-a243-dc871eef8e8c" containerName="nova-scheduler-scheduler" containerID="cri-o://07ed647e4210a85928967161b01d1a619e42389af3b1ff7678ac084276443c77" gracePeriod=30 Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.315415 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.315827 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6a61e6da-dded-4d12-b4f2-1907f0647817" containerName="nova-metadata-log" containerID="cri-o://3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df" gracePeriod=30 Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.315919 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6a61e6da-dded-4d12-b4f2-1907f0647817" containerName="nova-metadata-metadata" containerID="cri-o://ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170" gracePeriod=30 Dec 06 07:19:12 crc kubenswrapper[4945]: I1206 07:19:12.974725 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.028701 4945 generic.go:334] "Generic (PLEG): container finished" podID="6a61e6da-dded-4d12-b4f2-1907f0647817" containerID="ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170" exitCode=0 Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.028742 4945 generic.go:334] "Generic (PLEG): container finished" podID="6a61e6da-dded-4d12-b4f2-1907f0647817" containerID="3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df" exitCode=143 Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.028795 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a61e6da-dded-4d12-b4f2-1907f0647817","Type":"ContainerDied","Data":"ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170"} Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.028832 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a61e6da-dded-4d12-b4f2-1907f0647817","Type":"ContainerDied","Data":"3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df"} Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.028846 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a61e6da-dded-4d12-b4f2-1907f0647817","Type":"ContainerDied","Data":"c4660b40579cb0907fa70d1e8612c3403244e3c8bd277554990c3bd23e68eaf0"} Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.028863 4945 scope.go:117] "RemoveContainer" containerID="ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.029021 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.040712 4945 generic.go:334] "Generic (PLEG): container finished" podID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerID="a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4" exitCode=143 Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.040925 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0aa39339-6d1c-43db-87dd-ae3da33013f1","Type":"ContainerDied","Data":"a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4"} Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.076425 4945 scope.go:117] "RemoveContainer" containerID="3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.114561 4945 scope.go:117] "RemoveContainer" containerID="ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170" Dec 06 07:19:13 crc kubenswrapper[4945]: E1206 07:19:13.115206 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170\": container with ID starting with ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170 not found: ID does not exist" containerID="ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.115253 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170"} err="failed to get container status \"ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170\": rpc error: code = NotFound desc = could not find container \"ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170\": container with ID starting with ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170 not found: ID does not exist" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.115301 4945 scope.go:117] "RemoveContainer" containerID="3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df" Dec 06 07:19:13 crc kubenswrapper[4945]: E1206 07:19:13.115556 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df\": container with ID starting with 3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df not found: ID does not exist" containerID="3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.115586 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df"} err="failed to get container status \"3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df\": rpc error: code = NotFound desc = could not find container \"3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df\": container with ID starting with 3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df not found: ID does not exist" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.115606 4945 scope.go:117] "RemoveContainer" containerID="ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.115832 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170"} err="failed to get container status \"ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170\": rpc error: code = NotFound desc = could not find container \"ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170\": container with ID starting with ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170 not found: ID does not exist" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.115860 4945 scope.go:117] "RemoveContainer" containerID="3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.116108 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df"} err="failed to get container status \"3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df\": rpc error: code = NotFound desc = could not find container \"3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df\": container with ID starting with 3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df not found: ID does not exist" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.136054 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-nova-metadata-tls-certs\") pod \"6a61e6da-dded-4d12-b4f2-1907f0647817\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.136176 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a61e6da-dded-4d12-b4f2-1907f0647817-logs\") pod \"6a61e6da-dded-4d12-b4f2-1907f0647817\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.136223 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwmzb\" (UniqueName: \"kubernetes.io/projected/6a61e6da-dded-4d12-b4f2-1907f0647817-kube-api-access-rwmzb\") pod \"6a61e6da-dded-4d12-b4f2-1907f0647817\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.136331 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-config-data\") pod \"6a61e6da-dded-4d12-b4f2-1907f0647817\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.136379 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-combined-ca-bundle\") pod \"6a61e6da-dded-4d12-b4f2-1907f0647817\" (UID: \"6a61e6da-dded-4d12-b4f2-1907f0647817\") " Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.137262 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a61e6da-dded-4d12-b4f2-1907f0647817-logs" (OuterVolumeSpecName: "logs") pod "6a61e6da-dded-4d12-b4f2-1907f0647817" (UID: "6a61e6da-dded-4d12-b4f2-1907f0647817"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.144607 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a61e6da-dded-4d12-b4f2-1907f0647817-kube-api-access-rwmzb" (OuterVolumeSpecName: "kube-api-access-rwmzb") pod "6a61e6da-dded-4d12-b4f2-1907f0647817" (UID: "6a61e6da-dded-4d12-b4f2-1907f0647817"). InnerVolumeSpecName "kube-api-access-rwmzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.180820 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-config-data" (OuterVolumeSpecName: "config-data") pod "6a61e6da-dded-4d12-b4f2-1907f0647817" (UID: "6a61e6da-dded-4d12-b4f2-1907f0647817"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.184039 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a61e6da-dded-4d12-b4f2-1907f0647817" (UID: "6a61e6da-dded-4d12-b4f2-1907f0647817"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.202486 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6a61e6da-dded-4d12-b4f2-1907f0647817" (UID: "6a61e6da-dded-4d12-b4f2-1907f0647817"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.239256 4945 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.239321 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a61e6da-dded-4d12-b4f2-1907f0647817-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.239336 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwmzb\" (UniqueName: \"kubernetes.io/projected/6a61e6da-dded-4d12-b4f2-1907f0647817-kube-api-access-rwmzb\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.239348 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.239361 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a61e6da-dded-4d12-b4f2-1907f0647817-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:13 crc kubenswrapper[4945]: E1206 07:19:13.356881 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="07ed647e4210a85928967161b01d1a619e42389af3b1ff7678ac084276443c77" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 07:19:13 crc kubenswrapper[4945]: E1206 07:19:13.361900 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="07ed647e4210a85928967161b01d1a619e42389af3b1ff7678ac084276443c77" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 07:19:13 crc kubenswrapper[4945]: E1206 07:19:13.364223 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="07ed647e4210a85928967161b01d1a619e42389af3b1ff7678ac084276443c77" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 07:19:13 crc kubenswrapper[4945]: E1206 07:19:13.364333 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a3edb3c2-9c23-47e2-a243-dc871eef8e8c" containerName="nova-scheduler-scheduler" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.372447 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.386862 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.403726 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:13 crc kubenswrapper[4945]: E1206 07:19:13.404306 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a61e6da-dded-4d12-b4f2-1907f0647817" containerName="nova-metadata-log" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.404330 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a61e6da-dded-4d12-b4f2-1907f0647817" containerName="nova-metadata-log" Dec 06 07:19:13 crc kubenswrapper[4945]: E1206 07:19:13.404355 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5afc31d-67d1-4faf-998c-97b660f3d510" containerName="nova-manage" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.404363 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5afc31d-67d1-4faf-998c-97b660f3d510" containerName="nova-manage" Dec 06 07:19:13 crc kubenswrapper[4945]: E1206 07:19:13.404375 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a61e6da-dded-4d12-b4f2-1907f0647817" containerName="nova-metadata-metadata" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.404382 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a61e6da-dded-4d12-b4f2-1907f0647817" containerName="nova-metadata-metadata" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.404623 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a61e6da-dded-4d12-b4f2-1907f0647817" containerName="nova-metadata-log" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.404640 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a61e6da-dded-4d12-b4f2-1907f0647817" containerName="nova-metadata-metadata" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.404657 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5afc31d-67d1-4faf-998c-97b660f3d510" containerName="nova-manage" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.405991 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.412201 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.412847 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.443655 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.555372 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.555456 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.555499 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-config-data\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.555547 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-logs\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.555568 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f4gb\" (UniqueName: \"kubernetes.io/projected/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-kube-api-access-7f4gb\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.604113 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.657691 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.657771 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.657814 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-config-data\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.657861 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-logs\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.657879 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f4gb\" (UniqueName: \"kubernetes.io/projected/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-kube-api-access-7f4gb\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.670292 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-logs\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.675921 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.676679 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.686130 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-config-data\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.694009 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f4gb\" (UniqueName: \"kubernetes.io/projected/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-kube-api-access-7f4gb\") pod \"nova-metadata-0\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " pod="openstack/nova-metadata-0" Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.712918 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-8svps"] Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.713199 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" podUID="c6ac7f2c-e854-415e-9479-f808761d0a86" containerName="dnsmasq-dns" containerID="cri-o://97a7cb8a000b2b0ca98052a7ec3df67c8c15d0de75ce40f5229d9b011bb3c1b0" gracePeriod=10 Dec 06 07:19:13 crc kubenswrapper[4945]: I1206 07:19:13.858759 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.067553 4945 generic.go:334] "Generic (PLEG): container finished" podID="a3edb3c2-9c23-47e2-a243-dc871eef8e8c" containerID="07ed647e4210a85928967161b01d1a619e42389af3b1ff7678ac084276443c77" exitCode=0 Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.067627 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a3edb3c2-9c23-47e2-a243-dc871eef8e8c","Type":"ContainerDied","Data":"07ed647e4210a85928967161b01d1a619e42389af3b1ff7678ac084276443c77"} Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.070860 4945 generic.go:334] "Generic (PLEG): container finished" podID="b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b" containerID="0f377c0a02117dc7052bb48a4af37d61f3095dd82d7bde891095afbd858d61ee" exitCode=0 Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.070952 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kd5fs" event={"ID":"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b","Type":"ContainerDied","Data":"0f377c0a02117dc7052bb48a4af37d61f3095dd82d7bde891095afbd858d61ee"} Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.074388 4945 generic.go:334] "Generic (PLEG): container finished" podID="c6ac7f2c-e854-415e-9479-f808761d0a86" containerID="97a7cb8a000b2b0ca98052a7ec3df67c8c15d0de75ce40f5229d9b011bb3c1b0" exitCode=0 Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.074457 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" event={"ID":"c6ac7f2c-e854-415e-9479-f808761d0a86","Type":"ContainerDied","Data":"97a7cb8a000b2b0ca98052a7ec3df67c8c15d0de75ce40f5229d9b011bb3c1b0"} Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.284109 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.371070 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-svc\") pod \"c6ac7f2c-e854-415e-9479-f808761d0a86\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.371196 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-config\") pod \"c6ac7f2c-e854-415e-9479-f808761d0a86\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.371363 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-nb\") pod \"c6ac7f2c-e854-415e-9479-f808761d0a86\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.371550 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkpwn\" (UniqueName: \"kubernetes.io/projected/c6ac7f2c-e854-415e-9479-f808761d0a86-kube-api-access-pkpwn\") pod \"c6ac7f2c-e854-415e-9479-f808761d0a86\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.371681 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-sb\") pod \"c6ac7f2c-e854-415e-9479-f808761d0a86\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.371760 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-swift-storage-0\") pod \"c6ac7f2c-e854-415e-9479-f808761d0a86\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.379733 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6ac7f2c-e854-415e-9479-f808761d0a86-kube-api-access-pkpwn" (OuterVolumeSpecName: "kube-api-access-pkpwn") pod "c6ac7f2c-e854-415e-9479-f808761d0a86" (UID: "c6ac7f2c-e854-415e-9479-f808761d0a86"). InnerVolumeSpecName "kube-api-access-pkpwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.469800 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c6ac7f2c-e854-415e-9479-f808761d0a86" (UID: "c6ac7f2c-e854-415e-9479-f808761d0a86"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.476733 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c6ac7f2c-e854-415e-9479-f808761d0a86" (UID: "c6ac7f2c-e854-415e-9479-f808761d0a86"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.477472 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-svc\") pod \"c6ac7f2c-e854-415e-9479-f808761d0a86\" (UID: \"c6ac7f2c-e854-415e-9479-f808761d0a86\") " Dec 06 07:19:14 crc kubenswrapper[4945]: W1206 07:19:14.477561 4945 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c6ac7f2c-e854-415e-9479-f808761d0a86/volumes/kubernetes.io~configmap/dns-svc Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.477575 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c6ac7f2c-e854-415e-9479-f808761d0a86" (UID: "c6ac7f2c-e854-415e-9479-f808761d0a86"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.478347 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.478370 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkpwn\" (UniqueName: \"kubernetes.io/projected/c6ac7f2c-e854-415e-9479-f808761d0a86-kube-api-access-pkpwn\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.478386 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.497481 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c6ac7f2c-e854-415e-9479-f808761d0a86" (UID: "c6ac7f2c-e854-415e-9479-f808761d0a86"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.497481 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-config" (OuterVolumeSpecName: "config") pod "c6ac7f2c-e854-415e-9479-f808761d0a86" (UID: "c6ac7f2c-e854-415e-9479-f808761d0a86"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.503828 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c6ac7f2c-e854-415e-9479-f808761d0a86" (UID: "c6ac7f2c-e854-415e-9479-f808761d0a86"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.529425 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.580755 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.580810 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.580826 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6ac7f2c-e854-415e-9479-f808761d0a86-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.610607 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.682608 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffq46\" (UniqueName: \"kubernetes.io/projected/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-kube-api-access-ffq46\") pod \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.684000 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-config-data\") pod \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.684182 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-combined-ca-bundle\") pod \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\" (UID: \"a3edb3c2-9c23-47e2-a243-dc871eef8e8c\") " Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.687113 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-kube-api-access-ffq46" (OuterVolumeSpecName: "kube-api-access-ffq46") pod "a3edb3c2-9c23-47e2-a243-dc871eef8e8c" (UID: "a3edb3c2-9c23-47e2-a243-dc871eef8e8c"). InnerVolumeSpecName "kube-api-access-ffq46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.714477 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-config-data" (OuterVolumeSpecName: "config-data") pod "a3edb3c2-9c23-47e2-a243-dc871eef8e8c" (UID: "a3edb3c2-9c23-47e2-a243-dc871eef8e8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.719368 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3edb3c2-9c23-47e2-a243-dc871eef8e8c" (UID: "a3edb3c2-9c23-47e2-a243-dc871eef8e8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.787367 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffq46\" (UniqueName: \"kubernetes.io/projected/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-kube-api-access-ffq46\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.787408 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.787420 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3edb3c2-9c23-47e2-a243-dc871eef8e8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:14 crc kubenswrapper[4945]: I1206 07:19:14.967384 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a61e6da-dded-4d12-b4f2-1907f0647817" path="/var/lib/kubelet/pods/6a61e6da-dded-4d12-b4f2-1907f0647817/volumes" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.087976 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46","Type":"ContainerStarted","Data":"a3a38b3d67d1d5ebdd251a90a3d62bade79296bcdc1f39dd668b59634f5e68f5"} Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.088020 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46","Type":"ContainerStarted","Data":"b1fda35a0d6bd8c3f59f17fe7e7d7d2a2d9557c3711c2cfd122374a43bd94380"} Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.088029 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46","Type":"ContainerStarted","Data":"25eac9f656416b22c95285047781d4afc7c4f7fe46abfaf0dc6618285ac0eb9c"} Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.091244 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" event={"ID":"c6ac7f2c-e854-415e-9479-f808761d0a86","Type":"ContainerDied","Data":"ffa08229dd0758d42105d6513d2c2a445816fc8a39657db60371062ecf97b6f8"} Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.091310 4945 scope.go:117] "RemoveContainer" containerID="97a7cb8a000b2b0ca98052a7ec3df67c8c15d0de75ce40f5229d9b011bb3c1b0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.091425 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.095147 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.095378 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a3edb3c2-9c23-47e2-a243-dc871eef8e8c","Type":"ContainerDied","Data":"23f1fe5170291dd699701a9d8a8b954d4d5504453e28500637c032647607a23c"} Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.124296 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.124251715 podStartE2EDuration="2.124251715s" podCreationTimestamp="2025-12-06 07:19:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:15.116560447 +0000 UTC m=+1568.571421511" watchObservedRunningTime="2025-12-06 07:19:15.124251715 +0000 UTC m=+1568.579112759" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.140255 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.142391 4945 scope.go:117] "RemoveContainer" containerID="1484408d2bfa10e900e12b9b1944796a40e33648452c091db90f31342c12ad9f" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.152854 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.168373 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-8svps"] Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.177460 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:19:15 crc kubenswrapper[4945]: E1206 07:19:15.178072 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6ac7f2c-e854-415e-9479-f808761d0a86" containerName="dnsmasq-dns" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.178097 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6ac7f2c-e854-415e-9479-f808761d0a86" containerName="dnsmasq-dns" Dec 06 07:19:15 crc kubenswrapper[4945]: E1206 07:19:15.178145 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3edb3c2-9c23-47e2-a243-dc871eef8e8c" containerName="nova-scheduler-scheduler" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.178155 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3edb3c2-9c23-47e2-a243-dc871eef8e8c" containerName="nova-scheduler-scheduler" Dec 06 07:19:15 crc kubenswrapper[4945]: E1206 07:19:15.178169 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6ac7f2c-e854-415e-9479-f808761d0a86" containerName="init" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.178176 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6ac7f2c-e854-415e-9479-f808761d0a86" containerName="init" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.178437 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3edb3c2-9c23-47e2-a243-dc871eef8e8c" containerName="nova-scheduler-scheduler" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.178483 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6ac7f2c-e854-415e-9479-f808761d0a86" containerName="dnsmasq-dns" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.179354 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.181818 4945 scope.go:117] "RemoveContainer" containerID="07ed647e4210a85928967161b01d1a619e42389af3b1ff7678ac084276443c77" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.184688 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.203519 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-8svps"] Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.214914 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.298870 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcvcp\" (UniqueName: \"kubernetes.io/projected/5ade9957-3547-4e68-aab7-88256fddc367-kube-api-access-zcvcp\") pod \"nova-scheduler-0\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.298957 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.299005 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-config-data\") pod \"nova-scheduler-0\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.400679 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcvcp\" (UniqueName: \"kubernetes.io/projected/5ade9957-3547-4e68-aab7-88256fddc367-kube-api-access-zcvcp\") pod \"nova-scheduler-0\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.400771 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.400831 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-config-data\") pod \"nova-scheduler-0\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.407025 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.407874 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-config-data\") pod \"nova-scheduler-0\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.419530 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcvcp\" (UniqueName: \"kubernetes.io/projected/5ade9957-3547-4e68-aab7-88256fddc367-kube-api-access-zcvcp\") pod \"nova-scheduler-0\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.530009 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.671903 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.809762 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxqbf\" (UniqueName: \"kubernetes.io/projected/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-kube-api-access-zxqbf\") pod \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.809844 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-scripts\") pod \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.809916 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-config-data\") pod \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.810827 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-combined-ca-bundle\") pod \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\" (UID: \"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b\") " Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.816749 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-scripts" (OuterVolumeSpecName: "scripts") pod "b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b" (UID: "b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.832222 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-kube-api-access-zxqbf" (OuterVolumeSpecName: "kube-api-access-zxqbf") pod "b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b" (UID: "b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b"). InnerVolumeSpecName "kube-api-access-zxqbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.865597 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-config-data" (OuterVolumeSpecName: "config-data") pod "b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b" (UID: "b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.900395 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b" (UID: "b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.916698 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.916727 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxqbf\" (UniqueName: \"kubernetes.io/projected/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-kube-api-access-zxqbf\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.916737 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.916746 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:15 crc kubenswrapper[4945]: I1206 07:19:15.917070 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.018555 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-combined-ca-bundle\") pod \"0aa39339-6d1c-43db-87dd-ae3da33013f1\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.018752 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aa39339-6d1c-43db-87dd-ae3da33013f1-logs\") pod \"0aa39339-6d1c-43db-87dd-ae3da33013f1\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.018783 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-config-data\") pod \"0aa39339-6d1c-43db-87dd-ae3da33013f1\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.018924 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szb9b\" (UniqueName: \"kubernetes.io/projected/0aa39339-6d1c-43db-87dd-ae3da33013f1-kube-api-access-szb9b\") pod \"0aa39339-6d1c-43db-87dd-ae3da33013f1\" (UID: \"0aa39339-6d1c-43db-87dd-ae3da33013f1\") " Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.019443 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aa39339-6d1c-43db-87dd-ae3da33013f1-logs" (OuterVolumeSpecName: "logs") pod "0aa39339-6d1c-43db-87dd-ae3da33013f1" (UID: "0aa39339-6d1c-43db-87dd-ae3da33013f1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.019867 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aa39339-6d1c-43db-87dd-ae3da33013f1-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.023209 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aa39339-6d1c-43db-87dd-ae3da33013f1-kube-api-access-szb9b" (OuterVolumeSpecName: "kube-api-access-szb9b") pod "0aa39339-6d1c-43db-87dd-ae3da33013f1" (UID: "0aa39339-6d1c-43db-87dd-ae3da33013f1"). InnerVolumeSpecName "kube-api-access-szb9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.051139 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-config-data" (OuterVolumeSpecName: "config-data") pod "0aa39339-6d1c-43db-87dd-ae3da33013f1" (UID: "0aa39339-6d1c-43db-87dd-ae3da33013f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.054106 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0aa39339-6d1c-43db-87dd-ae3da33013f1" (UID: "0aa39339-6d1c-43db-87dd-ae3da33013f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.122766 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.122812 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szb9b\" (UniqueName: \"kubernetes.io/projected/0aa39339-6d1c-43db-87dd-ae3da33013f1-kube-api-access-szb9b\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.122828 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aa39339-6d1c-43db-87dd-ae3da33013f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.132638 4945 generic.go:334] "Generic (PLEG): container finished" podID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerID="087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718" exitCode=0 Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.132732 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0aa39339-6d1c-43db-87dd-ae3da33013f1","Type":"ContainerDied","Data":"087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718"} Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.132769 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0aa39339-6d1c-43db-87dd-ae3da33013f1","Type":"ContainerDied","Data":"205eeeedd1b5da38fa36f0f645db3488e1bd859608ef97d33991a4c6dd8b0abf"} Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.132792 4945 scope.go:117] "RemoveContainer" containerID="087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.132983 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.152091 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.152683 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kd5fs" event={"ID":"b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b","Type":"ContainerDied","Data":"e2789f73a3945a75266c91c64626f326df5fbcf70337bc683825b5f7a932c687"} Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.152720 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2789f73a3945a75266c91c64626f326df5fbcf70337bc683825b5f7a932c687" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.152821 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kd5fs" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.183206 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 07:19:16 crc kubenswrapper[4945]: E1206 07:19:16.184102 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerName="nova-api-log" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.184124 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerName="nova-api-log" Dec 06 07:19:16 crc kubenswrapper[4945]: E1206 07:19:16.184146 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b" containerName="nova-cell1-conductor-db-sync" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.184153 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b" containerName="nova-cell1-conductor-db-sync" Dec 06 07:19:16 crc kubenswrapper[4945]: E1206 07:19:16.184169 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerName="nova-api-api" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.184175 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerName="nova-api-api" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.184454 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerName="nova-api-log" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.184485 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b" containerName="nova-cell1-conductor-db-sync" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.184502 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" containerName="nova-api-api" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.185984 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.195043 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.232779 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.254717 4945 scope.go:117] "RemoveContainer" containerID="a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.254838 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.272341 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.283956 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.285829 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.288489 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.298001 4945 scope.go:117] "RemoveContainer" containerID="087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718" Dec 06 07:19:16 crc kubenswrapper[4945]: E1206 07:19:16.301361 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718\": container with ID starting with 087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718 not found: ID does not exist" containerID="087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.301418 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718"} err="failed to get container status \"087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718\": rpc error: code = NotFound desc = could not find container \"087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718\": container with ID starting with 087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718 not found: ID does not exist" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.301448 4945 scope.go:117] "RemoveContainer" containerID="a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4" Dec 06 07:19:16 crc kubenswrapper[4945]: E1206 07:19:16.304169 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4\": container with ID starting with a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4 not found: ID does not exist" containerID="a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.304259 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4"} err="failed to get container status \"a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4\": rpc error: code = NotFound desc = could not find container \"a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4\": container with ID starting with a62355eb15e4dfdcd0f56677625ca490e7f5a1d9c0b76d7f0ce67598766307c4 not found: ID does not exist" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.307199 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.326018 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.326116 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt2dc\" (UniqueName: \"kubernetes.io/projected/22eb2234-b14b-474c-b1bf-cf726acb064b-kube-api-access-jt2dc\") pod \"nova-cell1-conductor-0\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.326245 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.428453 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-logs\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.428787 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.428879 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-config-data\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.428962 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7k8x\" (UniqueName: \"kubernetes.io/projected/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-kube-api-access-n7k8x\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.429085 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.429217 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.429350 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt2dc\" (UniqueName: \"kubernetes.io/projected/22eb2234-b14b-474c-b1bf-cf726acb064b-kube-api-access-jt2dc\") pod \"nova-cell1-conductor-0\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.434452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.434941 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.447520 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt2dc\" (UniqueName: \"kubernetes.io/projected/22eb2234-b14b-474c-b1bf-cf726acb064b-kube-api-access-jt2dc\") pod \"nova-cell1-conductor-0\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.531310 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.531485 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-logs\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.531563 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-config-data\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.531582 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7k8x\" (UniqueName: \"kubernetes.io/projected/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-kube-api-access-n7k8x\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.532493 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-logs\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.536327 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.537307 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-config-data\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.537707 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.553702 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7k8x\" (UniqueName: \"kubernetes.io/projected/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-kube-api-access-n7k8x\") pod \"nova-api-0\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.614914 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.966338 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aa39339-6d1c-43db-87dd-ae3da33013f1" path="/var/lib/kubelet/pods/0aa39339-6d1c-43db-87dd-ae3da33013f1/volumes" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.967231 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3edb3c2-9c23-47e2-a243-dc871eef8e8c" path="/var/lib/kubelet/pods/a3edb3c2-9c23-47e2-a243-dc871eef8e8c/volumes" Dec 06 07:19:16 crc kubenswrapper[4945]: I1206 07:19:16.967780 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6ac7f2c-e854-415e-9479-f808761d0a86" path="/var/lib/kubelet/pods/c6ac7f2c-e854-415e-9479-f808761d0a86/volumes" Dec 06 07:19:17 crc kubenswrapper[4945]: I1206 07:19:17.038770 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 07:19:17 crc kubenswrapper[4945]: I1206 07:19:17.165760 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:17 crc kubenswrapper[4945]: I1206 07:19:17.166795 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"22eb2234-b14b-474c-b1bf-cf726acb064b","Type":"ContainerStarted","Data":"7f7a222718742cfc4420ef8b9284d3a809008a1d7a2ec4a2c47426e38684a78a"} Dec 06 07:19:17 crc kubenswrapper[4945]: I1206 07:19:17.174356 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5ade9957-3547-4e68-aab7-88256fddc367","Type":"ContainerStarted","Data":"8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6"} Dec 06 07:19:17 crc kubenswrapper[4945]: I1206 07:19:17.174402 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5ade9957-3547-4e68-aab7-88256fddc367","Type":"ContainerStarted","Data":"3ad6fe25b29a96b5212df8f9f2a163d388981ee36727e1d149e826c49b34b10b"} Dec 06 07:19:17 crc kubenswrapper[4945]: I1206 07:19:17.202895 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.20284372 podStartE2EDuration="2.20284372s" podCreationTimestamp="2025-12-06 07:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:17.197516406 +0000 UTC m=+1570.652377470" watchObservedRunningTime="2025-12-06 07:19:17.20284372 +0000 UTC m=+1570.657704784" Dec 06 07:19:18 crc kubenswrapper[4945]: I1206 07:19:18.189207 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"22eb2234-b14b-474c-b1bf-cf726acb064b","Type":"ContainerStarted","Data":"51e0e29bf34bea465946d1ff2713e5af788a26b0c2039d1356719fdaad5afb73"} Dec 06 07:19:18 crc kubenswrapper[4945]: I1206 07:19:18.189545 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:18 crc kubenswrapper[4945]: I1206 07:19:18.192384 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77","Type":"ContainerStarted","Data":"5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198"} Dec 06 07:19:18 crc kubenswrapper[4945]: I1206 07:19:18.192425 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77","Type":"ContainerStarted","Data":"87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb"} Dec 06 07:19:18 crc kubenswrapper[4945]: I1206 07:19:18.192439 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77","Type":"ContainerStarted","Data":"7458e66210566c0c3456e4a0114836523e2b8a892ab00ed93a69f23cdcbec1b7"} Dec 06 07:19:18 crc kubenswrapper[4945]: I1206 07:19:18.227562 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.227529332 podStartE2EDuration="2.227529332s" podCreationTimestamp="2025-12-06 07:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:18.216093333 +0000 UTC m=+1571.670954407" watchObservedRunningTime="2025-12-06 07:19:18.227529332 +0000 UTC m=+1571.682390376" Dec 06 07:19:18 crc kubenswrapper[4945]: I1206 07:19:18.258312 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.2582607230000002 podStartE2EDuration="2.258260723s" podCreationTimestamp="2025-12-06 07:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:18.241822299 +0000 UTC m=+1571.696683353" watchObservedRunningTime="2025-12-06 07:19:18.258260723 +0000 UTC m=+1571.713121767" Dec 06 07:19:18 crc kubenswrapper[4945]: I1206 07:19:18.859320 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 07:19:18 crc kubenswrapper[4945]: I1206 07:19:18.859388 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 07:19:19 crc kubenswrapper[4945]: I1206 07:19:19.077956 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77d8c9c7-8svps" podUID="c6ac7f2c-e854-415e-9479-f808761d0a86" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.160:5353: i/o timeout" Dec 06 07:19:20 crc kubenswrapper[4945]: I1206 07:19:20.531244 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 07:19:21 crc kubenswrapper[4945]: I1206 07:19:21.120173 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:21 crc kubenswrapper[4945]: I1206 07:19:21.176974 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:21 crc kubenswrapper[4945]: I1206 07:19:21.355456 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vxm6x"] Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.231939 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vxm6x" podUID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerName="registry-server" containerID="cri-o://07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa" gracePeriod=2 Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.760505 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.869244 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vb4l\" (UniqueName: \"kubernetes.io/projected/113b0590-9eee-4220-b7c9-44fcb69f250f-kube-api-access-7vb4l\") pod \"113b0590-9eee-4220-b7c9-44fcb69f250f\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.869447 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-catalog-content\") pod \"113b0590-9eee-4220-b7c9-44fcb69f250f\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.869600 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-utilities\") pod \"113b0590-9eee-4220-b7c9-44fcb69f250f\" (UID: \"113b0590-9eee-4220-b7c9-44fcb69f250f\") " Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.870192 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-utilities" (OuterVolumeSpecName: "utilities") pod "113b0590-9eee-4220-b7c9-44fcb69f250f" (UID: "113b0590-9eee-4220-b7c9-44fcb69f250f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.874942 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/113b0590-9eee-4220-b7c9-44fcb69f250f-kube-api-access-7vb4l" (OuterVolumeSpecName: "kube-api-access-7vb4l") pod "113b0590-9eee-4220-b7c9-44fcb69f250f" (UID: "113b0590-9eee-4220-b7c9-44fcb69f250f"). InnerVolumeSpecName "kube-api-access-7vb4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.940454 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "113b0590-9eee-4220-b7c9-44fcb69f250f" (UID: "113b0590-9eee-4220-b7c9-44fcb69f250f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.972447 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vb4l\" (UniqueName: \"kubernetes.io/projected/113b0590-9eee-4220-b7c9-44fcb69f250f-kube-api-access-7vb4l\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.972487 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:22 crc kubenswrapper[4945]: I1206 07:19:22.972499 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/113b0590-9eee-4220-b7c9-44fcb69f250f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.244092 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vxm6x" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.244214 4945 generic.go:334] "Generic (PLEG): container finished" podID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerID="07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa" exitCode=0 Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.245450 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vxm6x" event={"ID":"113b0590-9eee-4220-b7c9-44fcb69f250f","Type":"ContainerDied","Data":"07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa"} Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.245626 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vxm6x" event={"ID":"113b0590-9eee-4220-b7c9-44fcb69f250f","Type":"ContainerDied","Data":"55f4d5300e414665103682964b766dae6291dd9175b1e421d8e55f3880d04f22"} Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.245685 4945 scope.go:117] "RemoveContainer" containerID="07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.274808 4945 scope.go:117] "RemoveContainer" containerID="8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.274892 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vxm6x"] Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.284601 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vxm6x"] Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.306049 4945 scope.go:117] "RemoveContainer" containerID="40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.350943 4945 scope.go:117] "RemoveContainer" containerID="07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa" Dec 06 07:19:23 crc kubenswrapper[4945]: E1206 07:19:23.351703 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa\": container with ID starting with 07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa not found: ID does not exist" containerID="07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.351761 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa"} err="failed to get container status \"07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa\": rpc error: code = NotFound desc = could not find container \"07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa\": container with ID starting with 07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa not found: ID does not exist" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.351803 4945 scope.go:117] "RemoveContainer" containerID="8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74" Dec 06 07:19:23 crc kubenswrapper[4945]: E1206 07:19:23.352204 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74\": container with ID starting with 8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74 not found: ID does not exist" containerID="8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.352255 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74"} err="failed to get container status \"8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74\": rpc error: code = NotFound desc = could not find container \"8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74\": container with ID starting with 8ad8ea8168184cb80491599efad8a511cca837b0f04715891ccb0aad1a28bd74 not found: ID does not exist" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.352303 4945 scope.go:117] "RemoveContainer" containerID="40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1" Dec 06 07:19:23 crc kubenswrapper[4945]: E1206 07:19:23.352706 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1\": container with ID starting with 40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1 not found: ID does not exist" containerID="40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.352735 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1"} err="failed to get container status \"40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1\": rpc error: code = NotFound desc = could not find container \"40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1\": container with ID starting with 40cf61542b606a6fca258a475db75f2e72b74b64e599aada8e5f98d2b874c5d1 not found: ID does not exist" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.859535 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 07:19:23 crc kubenswrapper[4945]: I1206 07:19:23.859902 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 07:19:24 crc kubenswrapper[4945]: I1206 07:19:24.874487 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:19:24 crc kubenswrapper[4945]: I1206 07:19:24.874487 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:19:24 crc kubenswrapper[4945]: I1206 07:19:24.971006 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="113b0590-9eee-4220-b7c9-44fcb69f250f" path="/var/lib/kubelet/pods/113b0590-9eee-4220-b7c9-44fcb69f250f/volumes" Dec 06 07:19:25 crc kubenswrapper[4945]: I1206 07:19:25.531358 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 07:19:25 crc kubenswrapper[4945]: I1206 07:19:25.567192 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 07:19:25 crc kubenswrapper[4945]: E1206 07:19:25.958375 4945 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/265743d5ea048a3c32fd8311d35127f0e47eb87f03afa71d0a3a26f1ceee84da/diff" to get inode usage: stat /var/lib/containers/storage/overlay/265743d5ea048a3c32fd8311d35127f0e47eb87f03afa71d0a3a26f1ceee84da/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-77d8c9c7-8svps_c6ac7f2c-e854-415e-9479-f808761d0a86/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-77d8c9c7-8svps_c6ac7f2c-e854-415e-9479-f808761d0a86/dnsmasq-dns/0.log: no such file or directory Dec 06 07:19:26 crc kubenswrapper[4945]: I1206 07:19:26.314770 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 07:19:26 crc kubenswrapper[4945]: I1206 07:19:26.572615 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 06 07:19:26 crc kubenswrapper[4945]: I1206 07:19:26.616191 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 07:19:26 crc kubenswrapper[4945]: I1206 07:19:26.618392 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 07:19:27 crc kubenswrapper[4945]: I1206 07:19:27.698550 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:19:27 crc kubenswrapper[4945]: I1206 07:19:27.698528 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:19:28 crc kubenswrapper[4945]: I1206 07:19:28.053801 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 07:19:33 crc kubenswrapper[4945]: I1206 07:19:33.865763 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 07:19:33 crc kubenswrapper[4945]: I1206 07:19:33.866466 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 07:19:33 crc kubenswrapper[4945]: I1206 07:19:33.872680 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 07:19:33 crc kubenswrapper[4945]: I1206 07:19:33.872871 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.005932 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hssjc"] Dec 06 07:19:34 crc kubenswrapper[4945]: E1206 07:19:34.018126 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerName="extract-utilities" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.018169 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerName="extract-utilities" Dec 06 07:19:34 crc kubenswrapper[4945]: E1206 07:19:34.018187 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerName="registry-server" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.018195 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerName="registry-server" Dec 06 07:19:34 crc kubenswrapper[4945]: E1206 07:19:34.018219 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerName="extract-content" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.018226 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerName="extract-content" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.018549 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="113b0590-9eee-4220-b7c9-44fcb69f250f" containerName="registry-server" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.020341 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.029362 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hssjc"] Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.137607 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-utilities\") pod \"redhat-marketplace-hssjc\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.137769 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzsnj\" (UniqueName: \"kubernetes.io/projected/05954cf8-6ddf-4455-9b02-8f75b07f6272-kube-api-access-bzsnj\") pod \"redhat-marketplace-hssjc\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.137792 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-catalog-content\") pod \"redhat-marketplace-hssjc\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.239703 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzsnj\" (UniqueName: \"kubernetes.io/projected/05954cf8-6ddf-4455-9b02-8f75b07f6272-kube-api-access-bzsnj\") pod \"redhat-marketplace-hssjc\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.240300 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-catalog-content\") pod \"redhat-marketplace-hssjc\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.240390 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-utilities\") pod \"redhat-marketplace-hssjc\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.240894 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-catalog-content\") pod \"redhat-marketplace-hssjc\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.240931 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-utilities\") pod \"redhat-marketplace-hssjc\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.264192 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzsnj\" (UniqueName: \"kubernetes.io/projected/05954cf8-6ddf-4455-9b02-8f75b07f6272-kube-api-access-bzsnj\") pod \"redhat-marketplace-hssjc\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.345256 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:34 crc kubenswrapper[4945]: I1206 07:19:34.927099 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hssjc"] Dec 06 07:19:35 crc kubenswrapper[4945]: I1206 07:19:35.394949 4945 generic.go:334] "Generic (PLEG): container finished" podID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerID="b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1" exitCode=0 Dec 06 07:19:35 crc kubenswrapper[4945]: I1206 07:19:35.395060 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hssjc" event={"ID":"05954cf8-6ddf-4455-9b02-8f75b07f6272","Type":"ContainerDied","Data":"b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1"} Dec 06 07:19:35 crc kubenswrapper[4945]: I1206 07:19:35.395544 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hssjc" event={"ID":"05954cf8-6ddf-4455-9b02-8f75b07f6272","Type":"ContainerStarted","Data":"1cd62d9941aa0fde8450059557570108ff35355f0b45a491e20e374b2c7b1a6a"} Dec 06 07:19:35 crc kubenswrapper[4945]: W1206 07:19:35.918503 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a61e6da_dded_4d12_b4f2_1907f0647817.slice/crio-c4660b40579cb0907fa70d1e8612c3403244e3c8bd277554990c3bd23e68eaf0 WatchSource:0}: Error finding container c4660b40579cb0907fa70d1e8612c3403244e3c8bd277554990c3bd23e68eaf0: Status 404 returned error can't find the container with id c4660b40579cb0907fa70d1e8612c3403244e3c8bd277554990c3bd23e68eaf0 Dec 06 07:19:35 crc kubenswrapper[4945]: W1206 07:19:35.920300 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod113b0590_9eee_4220_b7c9_44fcb69f250f.slice/crio-07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa.scope WatchSource:0}: Error finding container 07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa: Status 404 returned error can't find the container with id 07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa Dec 06 07:19:35 crc kubenswrapper[4945]: W1206 07:19:35.920738 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a61e6da_dded_4d12_b4f2_1907f0647817.slice/crio-3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df.scope WatchSource:0}: Error finding container 3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df: Status 404 returned error can't find the container with id 3d0f6c557b32b2f83986e1b7bd5893c681beb3cdd9f61b26467325833298a8df Dec 06 07:19:35 crc kubenswrapper[4945]: W1206 07:19:35.922245 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a61e6da_dded_4d12_b4f2_1907f0647817.slice/crio-ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170.scope WatchSource:0}: Error finding container ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170: Status 404 returned error can't find the container with id ddaac4b70581cfd886c820af30f001d9f67adb6add472c4346cf965418b6d170 Dec 06 07:19:36 crc kubenswrapper[4945]: E1206 07:19:36.277921 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8b7d3fd_fcf0_4f38_b2d1_31b883b0922b.slice/crio-conmon-0f377c0a02117dc7052bb48a4af37d61f3095dd82d7bde891095afbd858d61ee.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0aa39339_6d1c_43db_87dd_ae3da33013f1.slice/crio-087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8b7d3fd_fcf0_4f38_b2d1_31b883b0922b.slice/crio-e2789f73a3945a75266c91c64626f326df5fbcf70337bc683825b5f7a932c687\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0aa39339_6d1c_43db_87dd_ae3da33013f1.slice/crio-205eeeedd1b5da38fa36f0f645db3488e1bd859608ef97d33991a4c6dd8b0abf\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3edb3c2_9c23_47e2_a243_dc871eef8e8c.slice/crio-conmon-07ed647e4210a85928967161b01d1a619e42389af3b1ff7678ac084276443c77.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6ac7f2c_e854_415e_9479_f808761d0a86.slice/crio-97a7cb8a000b2b0ca98052a7ec3df67c8c15d0de75ce40f5229d9b011bb3c1b0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8b7d3fd_fcf0_4f38_b2d1_31b883b0922b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6ac7f2c_e854_415e_9479_f808761d0a86.slice/crio-conmon-97a7cb8a000b2b0ca98052a7ec3df67c8c15d0de75ce40f5229d9b011bb3c1b0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0aa39339_6d1c_43db_87dd_ae3da33013f1.slice/crio-conmon-087ea8e9f339cc9bfe4ba32439fffcb288d4cf20bc13a66f19c6e4e76a859718.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3edb3c2_9c23_47e2_a243_dc871eef8e8c.slice/crio-23f1fe5170291dd699701a9d8a8b954d4d5504453e28500637c032647607a23c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0aa39339_6d1c_43db_87dd_ae3da33013f1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3edb3c2_9c23_47e2_a243_dc871eef8e8c.slice/crio-07ed647e4210a85928967161b01d1a619e42389af3b1ff7678ac084276443c77.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6ac7f2c_e854_415e_9479_f808761d0a86.slice/crio-ffa08229dd0758d42105d6513d2c2a445816fc8a39657db60371062ecf97b6f8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod113b0590_9eee_4220_b7c9_44fcb69f250f.slice/crio-55f4d5300e414665103682964b766dae6291dd9175b1e421d8e55f3880d04f22\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6ac7f2c_e854_415e_9479_f808761d0a86.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod113b0590_9eee_4220_b7c9_44fcb69f250f.slice/crio-conmon-07fc39d5e13042b66911ae75291abf0c22e7ee5823cb095c9433ce0c57e32caa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3edb3c2_9c23_47e2_a243_dc871eef8e8c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod113b0590_9eee_4220_b7c9_44fcb69f250f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54d6c3b2_3db2_4400_958e_839a0f53cda0.slice/crio-conmon-0ab1417aa10f14854712f6287bc7ff98a77fec3e4030e9d198a868daeb2bd9f6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54d6c3b2_3db2_4400_958e_839a0f53cda0.slice/crio-0ab1417aa10f14854712f6287bc7ff98a77fec3e4030e9d198a868daeb2bd9f6.scope\": RecentStats: unable to find data in memory cache]" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.410328 4945 generic.go:334] "Generic (PLEG): container finished" podID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerID="46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c" exitCode=0 Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.410415 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hssjc" event={"ID":"05954cf8-6ddf-4455-9b02-8f75b07f6272","Type":"ContainerDied","Data":"46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c"} Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.415988 4945 generic.go:334] "Generic (PLEG): container finished" podID="54d6c3b2-3db2-4400-958e-839a0f53cda0" containerID="0ab1417aa10f14854712f6287bc7ff98a77fec3e4030e9d198a868daeb2bd9f6" exitCode=137 Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.416037 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"54d6c3b2-3db2-4400-958e-839a0f53cda0","Type":"ContainerDied","Data":"0ab1417aa10f14854712f6287bc7ff98a77fec3e4030e9d198a868daeb2bd9f6"} Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.416075 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"54d6c3b2-3db2-4400-958e-839a0f53cda0","Type":"ContainerDied","Data":"0ed337f4ef64726823d490c493b0dd3c7f12e9aa048c8332fe01a3bf01c73a1e"} Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.416088 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ed337f4ef64726823d490c493b0dd3c7f12e9aa048c8332fe01a3bf01c73a1e" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.482752 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.591756 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffxkj\" (UniqueName: \"kubernetes.io/projected/54d6c3b2-3db2-4400-958e-839a0f53cda0-kube-api-access-ffxkj\") pod \"54d6c3b2-3db2-4400-958e-839a0f53cda0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.591874 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-combined-ca-bundle\") pod \"54d6c3b2-3db2-4400-958e-839a0f53cda0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.591921 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-config-data\") pod \"54d6c3b2-3db2-4400-958e-839a0f53cda0\" (UID: \"54d6c3b2-3db2-4400-958e-839a0f53cda0\") " Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.609775 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54d6c3b2-3db2-4400-958e-839a0f53cda0-kube-api-access-ffxkj" (OuterVolumeSpecName: "kube-api-access-ffxkj") pod "54d6c3b2-3db2-4400-958e-839a0f53cda0" (UID: "54d6c3b2-3db2-4400-958e-839a0f53cda0"). InnerVolumeSpecName "kube-api-access-ffxkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.625219 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54d6c3b2-3db2-4400-958e-839a0f53cda0" (UID: "54d6c3b2-3db2-4400-958e-839a0f53cda0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.626702 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.626868 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-config-data" (OuterVolumeSpecName: "config-data") pod "54d6c3b2-3db2-4400-958e-839a0f53cda0" (UID: "54d6c3b2-3db2-4400-958e-839a0f53cda0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.627659 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.627688 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.630756 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.694520 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffxkj\" (UniqueName: \"kubernetes.io/projected/54d6c3b2-3db2-4400-958e-839a0f53cda0-kube-api-access-ffxkj\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.694852 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:36 crc kubenswrapper[4945]: I1206 07:19:36.694868 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d6c3b2-3db2-4400-958e-839a0f53cda0-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.429186 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.429201 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hssjc" event={"ID":"05954cf8-6ddf-4455-9b02-8f75b07f6272","Type":"ContainerStarted","Data":"6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f"} Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.430464 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.461136 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hssjc" podStartSLOduration=3.050761551 podStartE2EDuration="4.461111215s" podCreationTimestamp="2025-12-06 07:19:33 +0000 UTC" firstStartedPulling="2025-12-06 07:19:35.397420905 +0000 UTC m=+1588.852281949" lastFinishedPulling="2025-12-06 07:19:36.807770569 +0000 UTC m=+1590.262631613" observedRunningTime="2025-12-06 07:19:37.449707566 +0000 UTC m=+1590.904568620" watchObservedRunningTime="2025-12-06 07:19:37.461111215 +0000 UTC m=+1590.915972259" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.486230 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.494201 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.502778 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.644360 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:19:37 crc kubenswrapper[4945]: E1206 07:19:37.651664 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d6c3b2-3db2-4400-958e-839a0f53cda0" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.651702 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d6c3b2-3db2-4400-958e-839a0f53cda0" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.652140 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d6c3b2-3db2-4400-958e-839a0f53cda0" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.653155 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.659470 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.659547 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.678473 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.690249 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.744914 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.744988 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.745068 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w2f9\" (UniqueName: \"kubernetes.io/projected/3a31495b-7caf-4037-87b3-c0d7f29bf582-kube-api-access-5w2f9\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.745125 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.745391 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.847560 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.847761 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.847832 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.847858 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.847926 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w2f9\" (UniqueName: \"kubernetes.io/projected/3a31495b-7caf-4037-87b3-c0d7f29bf582-kube-api-access-5w2f9\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.853371 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.853696 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.854480 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.857224 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.894289 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-n747t"] Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.896312 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.896653 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w2f9\" (UniqueName: \"kubernetes.io/projected/3a31495b-7caf-4037-87b3-c0d7f29bf582-kube-api-access-5w2f9\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.923205 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-n747t"] Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.950494 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-nb\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.950554 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-sb\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.950587 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpch8\" (UniqueName: \"kubernetes.io/projected/384f1bb0-ff94-4fc4-b658-7e1a250945ae-kube-api-access-dpch8\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.950644 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-swift-storage-0\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.950835 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-svc\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.951862 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-config\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:37 crc kubenswrapper[4945]: I1206 07:19:37.997864 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.053897 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-swift-storage-0\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.054296 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-svc\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.054355 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-config\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.054412 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-nb\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.054432 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-sb\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.054460 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpch8\" (UniqueName: \"kubernetes.io/projected/384f1bb0-ff94-4fc4-b658-7e1a250945ae-kube-api-access-dpch8\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.054947 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-swift-storage-0\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.055260 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-svc\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.055326 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-config\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.058858 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-nb\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.059818 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-sb\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.080272 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpch8\" (UniqueName: \"kubernetes.io/projected/384f1bb0-ff94-4fc4-b658-7e1a250945ae-kube-api-access-dpch8\") pod \"dnsmasq-dns-cf4ff87b5-n747t\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.271387 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.545665 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.886065 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-n747t"] Dec 06 07:19:38 crc kubenswrapper[4945]: I1206 07:19:38.981014 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54d6c3b2-3db2-4400-958e-839a0f53cda0" path="/var/lib/kubelet/pods/54d6c3b2-3db2-4400-958e-839a0f53cda0/volumes" Dec 06 07:19:39 crc kubenswrapper[4945]: I1206 07:19:39.454219 4945 generic.go:334] "Generic (PLEG): container finished" podID="384f1bb0-ff94-4fc4-b658-7e1a250945ae" containerID="55421545d7e8b51ad480e6adea462be55656e1f3e06b69e287fb97a724e2a633" exitCode=0 Dec 06 07:19:39 crc kubenswrapper[4945]: I1206 07:19:39.454306 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" event={"ID":"384f1bb0-ff94-4fc4-b658-7e1a250945ae","Type":"ContainerDied","Data":"55421545d7e8b51ad480e6adea462be55656e1f3e06b69e287fb97a724e2a633"} Dec 06 07:19:39 crc kubenswrapper[4945]: I1206 07:19:39.454596 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" event={"ID":"384f1bb0-ff94-4fc4-b658-7e1a250945ae","Type":"ContainerStarted","Data":"c6628efe2ca405c74026daa5ade200e1299c7929321b142e8f16513a47a06e6c"} Dec 06 07:19:39 crc kubenswrapper[4945]: I1206 07:19:39.456679 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a31495b-7caf-4037-87b3-c0d7f29bf582","Type":"ContainerStarted","Data":"89f2192e8e4d5ac29a2fb860d1dd70435b3661183bcc6f14f8a20a90fab12155"} Dec 06 07:19:39 crc kubenswrapper[4945]: I1206 07:19:39.456718 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a31495b-7caf-4037-87b3-c0d7f29bf582","Type":"ContainerStarted","Data":"4d06076acad515ac5126efa1ff728044607ce8bccc093f1dec2556b021c8f8af"} Dec 06 07:19:39 crc kubenswrapper[4945]: I1206 07:19:39.509931 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.509911914 podStartE2EDuration="2.509911914s" podCreationTimestamp="2025-12-06 07:19:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:39.504645671 +0000 UTC m=+1592.959506715" watchObservedRunningTime="2025-12-06 07:19:39.509911914 +0000 UTC m=+1592.964772958" Dec 06 07:19:40 crc kubenswrapper[4945]: I1206 07:19:40.471028 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" event={"ID":"384f1bb0-ff94-4fc4-b658-7e1a250945ae","Type":"ContainerStarted","Data":"cf42f6e43b6eaa14b9d60d2d2162c0cfe2ef17791091ee3c53ee8ba8cb1d8339"} Dec 06 07:19:40 crc kubenswrapper[4945]: I1206 07:19:40.528165 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" podStartSLOduration=3.5281322299999998 podStartE2EDuration="3.52813223s" podCreationTimestamp="2025-12-06 07:19:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:40.511703385 +0000 UTC m=+1593.966564429" watchObservedRunningTime="2025-12-06 07:19:40.52813223 +0000 UTC m=+1593.982993284" Dec 06 07:19:40 crc kubenswrapper[4945]: I1206 07:19:40.624245 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:19:40 crc kubenswrapper[4945]: I1206 07:19:40.624785 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="ceilometer-central-agent" containerID="cri-o://b9af4bd5a62cda53dd725a20e1df6be731fdc58e780f4838726be7126c072f5a" gracePeriod=30 Dec 06 07:19:40 crc kubenswrapper[4945]: I1206 07:19:40.624838 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="ceilometer-notification-agent" containerID="cri-o://5ed82833aca583fbab32a527da4bef5b512b77c7c64397f7742dbe74a40921e4" gracePeriod=30 Dec 06 07:19:40 crc kubenswrapper[4945]: I1206 07:19:40.624798 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="proxy-httpd" containerID="cri-o://58a37447d6a7e82e60ac8bec54e4492001005996d111a5d73ac2f6fa2b4804b7" gracePeriod=30 Dec 06 07:19:40 crc kubenswrapper[4945]: I1206 07:19:40.625047 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="sg-core" containerID="cri-o://c36d1498eada2e39ab90b6f41041b808f19d09761fcb2090a1abd248c195a6c3" gracePeriod=30 Dec 06 07:19:40 crc kubenswrapper[4945]: I1206 07:19:40.699341 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:40 crc kubenswrapper[4945]: I1206 07:19:40.699837 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerName="nova-api-log" containerID="cri-o://87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb" gracePeriod=30 Dec 06 07:19:40 crc kubenswrapper[4945]: I1206 07:19:40.700527 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerName="nova-api-api" containerID="cri-o://5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198" gracePeriod=30 Dec 06 07:19:41 crc kubenswrapper[4945]: I1206 07:19:41.482769 4945 generic.go:334] "Generic (PLEG): container finished" podID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerID="58a37447d6a7e82e60ac8bec54e4492001005996d111a5d73ac2f6fa2b4804b7" exitCode=0 Dec 06 07:19:41 crc kubenswrapper[4945]: I1206 07:19:41.482814 4945 generic.go:334] "Generic (PLEG): container finished" podID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerID="c36d1498eada2e39ab90b6f41041b808f19d09761fcb2090a1abd248c195a6c3" exitCode=2 Dec 06 07:19:41 crc kubenswrapper[4945]: I1206 07:19:41.482826 4945 generic.go:334] "Generic (PLEG): container finished" podID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerID="b9af4bd5a62cda53dd725a20e1df6be731fdc58e780f4838726be7126c072f5a" exitCode=0 Dec 06 07:19:41 crc kubenswrapper[4945]: I1206 07:19:41.482847 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1eb7e6e-6348-4285-81ff-ae05307cc91e","Type":"ContainerDied","Data":"58a37447d6a7e82e60ac8bec54e4492001005996d111a5d73ac2f6fa2b4804b7"} Dec 06 07:19:41 crc kubenswrapper[4945]: I1206 07:19:41.482884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1eb7e6e-6348-4285-81ff-ae05307cc91e","Type":"ContainerDied","Data":"c36d1498eada2e39ab90b6f41041b808f19d09761fcb2090a1abd248c195a6c3"} Dec 06 07:19:41 crc kubenswrapper[4945]: I1206 07:19:41.482895 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1eb7e6e-6348-4285-81ff-ae05307cc91e","Type":"ContainerDied","Data":"b9af4bd5a62cda53dd725a20e1df6be731fdc58e780f4838726be7126c072f5a"} Dec 06 07:19:41 crc kubenswrapper[4945]: I1206 07:19:41.485123 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerID="87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb" exitCode=143 Dec 06 07:19:41 crc kubenswrapper[4945]: I1206 07:19:41.485227 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77","Type":"ContainerDied","Data":"87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb"} Dec 06 07:19:41 crc kubenswrapper[4945]: I1206 07:19:41.485450 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:42 crc kubenswrapper[4945]: I1206 07:19:42.998007 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:43 crc kubenswrapper[4945]: I1206 07:19:43.562309 4945 generic.go:334] "Generic (PLEG): container finished" podID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerID="5ed82833aca583fbab32a527da4bef5b512b77c7c64397f7742dbe74a40921e4" exitCode=0 Dec 06 07:19:43 crc kubenswrapper[4945]: I1206 07:19:43.562361 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1eb7e6e-6348-4285-81ff-ae05307cc91e","Type":"ContainerDied","Data":"5ed82833aca583fbab32a527da4bef5b512b77c7c64397f7742dbe74a40921e4"} Dec 06 07:19:43 crc kubenswrapper[4945]: I1206 07:19:43.968306 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.087834 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-log-httpd\") pod \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088134 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-combined-ca-bundle\") pod \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088187 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-run-httpd\") pod \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088228 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78bmq\" (UniqueName: \"kubernetes.io/projected/b1eb7e6e-6348-4285-81ff-ae05307cc91e-kube-api-access-78bmq\") pod \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088311 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-scripts\") pod \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088332 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-ceilometer-tls-certs\") pod \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088352 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-sg-core-conf-yaml\") pod \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088408 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-config-data\") pod \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\" (UID: \"b1eb7e6e-6348-4285-81ff-ae05307cc91e\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088406 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b1eb7e6e-6348-4285-81ff-ae05307cc91e" (UID: "b1eb7e6e-6348-4285-81ff-ae05307cc91e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088619 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b1eb7e6e-6348-4285-81ff-ae05307cc91e" (UID: "b1eb7e6e-6348-4285-81ff-ae05307cc91e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088880 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.088899 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1eb7e6e-6348-4285-81ff-ae05307cc91e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.117607 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1eb7e6e-6348-4285-81ff-ae05307cc91e-kube-api-access-78bmq" (OuterVolumeSpecName: "kube-api-access-78bmq") pod "b1eb7e6e-6348-4285-81ff-ae05307cc91e" (UID: "b1eb7e6e-6348-4285-81ff-ae05307cc91e"). InnerVolumeSpecName "kube-api-access-78bmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.119441 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-scripts" (OuterVolumeSpecName: "scripts") pod "b1eb7e6e-6348-4285-81ff-ae05307cc91e" (UID: "b1eb7e6e-6348-4285-81ff-ae05307cc91e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.152431 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b1eb7e6e-6348-4285-81ff-ae05307cc91e" (UID: "b1eb7e6e-6348-4285-81ff-ae05307cc91e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.166613 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b1eb7e6e-6348-4285-81ff-ae05307cc91e" (UID: "b1eb7e6e-6348-4285-81ff-ae05307cc91e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.191565 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78bmq\" (UniqueName: \"kubernetes.io/projected/b1eb7e6e-6348-4285-81ff-ae05307cc91e-kube-api-access-78bmq\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.191994 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.192254 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.192379 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.225523 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1eb7e6e-6348-4285-81ff-ae05307cc91e" (UID: "b1eb7e6e-6348-4285-81ff-ae05307cc91e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.279137 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-config-data" (OuterVolumeSpecName: "config-data") pod "b1eb7e6e-6348-4285-81ff-ae05307cc91e" (UID: "b1eb7e6e-6348-4285-81ff-ae05307cc91e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.297104 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.297149 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1eb7e6e-6348-4285-81ff-ae05307cc91e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.309507 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.347978 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.348374 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.398711 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-logs\") pod \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.398910 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-config-data\") pod \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.398931 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-combined-ca-bundle\") pod \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.399035 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7k8x\" (UniqueName: \"kubernetes.io/projected/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-kube-api-access-n7k8x\") pod \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\" (UID: \"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77\") " Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.400378 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-logs" (OuterVolumeSpecName: "logs") pod "3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" (UID: "3c020ac4-55e5-4e1e-ab07-2dae1fc62b77"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.407037 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.416677 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-kube-api-access-n7k8x" (OuterVolumeSpecName: "kube-api-access-n7k8x") pod "3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" (UID: "3c020ac4-55e5-4e1e-ab07-2dae1fc62b77"). InnerVolumeSpecName "kube-api-access-n7k8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.435789 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-config-data" (OuterVolumeSpecName: "config-data") pod "3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" (UID: "3c020ac4-55e5-4e1e-ab07-2dae1fc62b77"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.449585 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" (UID: "3c020ac4-55e5-4e1e-ab07-2dae1fc62b77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.501843 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.501881 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.501891 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7k8x\" (UniqueName: \"kubernetes.io/projected/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-kube-api-access-n7k8x\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.501901 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.575326 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1eb7e6e-6348-4285-81ff-ae05307cc91e","Type":"ContainerDied","Data":"7389dc1f6c51556cd21183504f03406664433f35936be7115037f2b47bade53b"} Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.575377 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.575391 4945 scope.go:117] "RemoveContainer" containerID="58a37447d6a7e82e60ac8bec54e4492001005996d111a5d73ac2f6fa2b4804b7" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.582947 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerID="5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198" exitCode=0 Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.583047 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.583055 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77","Type":"ContainerDied","Data":"5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198"} Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.583108 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c020ac4-55e5-4e1e-ab07-2dae1fc62b77","Type":"ContainerDied","Data":"7458e66210566c0c3456e4a0114836523e2b8a892ab00ed93a69f23cdcbec1b7"} Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.623222 4945 scope.go:117] "RemoveContainer" containerID="c36d1498eada2e39ab90b6f41041b808f19d09761fcb2090a1abd248c195a6c3" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.650465 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.665843 4945 scope.go:117] "RemoveContainer" containerID="5ed82833aca583fbab32a527da4bef5b512b77c7c64397f7742dbe74a40921e4" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.672482 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.707319 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.729545 4945 scope.go:117] "RemoveContainer" containerID="b9af4bd5a62cda53dd725a20e1df6be731fdc58e780f4838726be7126c072f5a" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.774909 4945 scope.go:117] "RemoveContainer" containerID="5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.775379 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.794882 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.803629 4945 scope.go:117] "RemoveContainer" containerID="87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.816186 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:19:44 crc kubenswrapper[4945]: E1206 07:19:44.817039 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="ceilometer-central-agent" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.817143 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="ceilometer-central-agent" Dec 06 07:19:44 crc kubenswrapper[4945]: E1206 07:19:44.817272 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="ceilometer-notification-agent" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.817410 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="ceilometer-notification-agent" Dec 06 07:19:44 crc kubenswrapper[4945]: E1206 07:19:44.817466 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="proxy-httpd" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.817672 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="proxy-httpd" Dec 06 07:19:44 crc kubenswrapper[4945]: E1206 07:19:44.817766 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="sg-core" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.817840 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="sg-core" Dec 06 07:19:44 crc kubenswrapper[4945]: E1206 07:19:44.817908 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerName="nova-api-log" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.818014 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerName="nova-api-log" Dec 06 07:19:44 crc kubenswrapper[4945]: E1206 07:19:44.818100 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerName="nova-api-api" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.818153 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerName="nova-api-api" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.818505 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerName="nova-api-api" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.818579 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="sg-core" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.818650 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" containerName="nova-api-log" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.818713 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="proxy-httpd" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.818774 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="ceilometer-notification-agent" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.818845 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" containerName="ceilometer-central-agent" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.820876 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.823935 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.827136 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.827197 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.829143 4945 scope.go:117] "RemoveContainer" containerID="5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.829532 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:44 crc kubenswrapper[4945]: E1206 07:19:44.829820 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198\": container with ID starting with 5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198 not found: ID does not exist" containerID="5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.829849 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198"} err="failed to get container status \"5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198\": rpc error: code = NotFound desc = could not find container \"5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198\": container with ID starting with 5bc8a05020ddfbcb0da3fddf22cae752a9bb6ed1dd82446303942ebc72b6b198 not found: ID does not exist" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.829870 4945 scope.go:117] "RemoveContainer" containerID="87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.832084 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:19:44 crc kubenswrapper[4945]: E1206 07:19:44.833572 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb\": container with ID starting with 87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb not found: ID does not exist" containerID="87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.833611 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb"} err="failed to get container status \"87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb\": rpc error: code = NotFound desc = could not find container \"87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb\": container with ID starting with 87aac1d9a91013bc209a4a8da05e07eb94978110f115cc125ce989d014ddcbbb not found: ID does not exist" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.834083 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.834272 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.834372 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.856249 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.864928 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.883157 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hssjc"] Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.915879 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg6rk\" (UniqueName: \"kubernetes.io/projected/65affa4a-2f58-4de2-9533-18e2d72af609-kube-api-access-jg6rk\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.916320 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-public-tls-certs\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.916472 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-scripts\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.916577 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-config-data\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.916661 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.916766 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.916890 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.917122 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65affa4a-2f58-4de2-9533-18e2d72af609-logs\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.917184 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.917229 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5wqh\" (UniqueName: \"kubernetes.io/projected/f8fbb8a7-4226-4b2f-9e95-621059e6283d-kube-api-access-d5wqh\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.917369 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-config-data\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.917420 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-log-httpd\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.917533 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-run-httpd\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.917636 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.965771 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c020ac4-55e5-4e1e-ab07-2dae1fc62b77" path="/var/lib/kubelet/pods/3c020ac4-55e5-4e1e-ab07-2dae1fc62b77/volumes" Dec 06 07:19:44 crc kubenswrapper[4945]: I1206 07:19:44.966797 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1eb7e6e-6348-4285-81ff-ae05307cc91e" path="/var/lib/kubelet/pods/b1eb7e6e-6348-4285-81ff-ae05307cc91e/volumes" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.019671 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65affa4a-2f58-4de2-9533-18e2d72af609-logs\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.019729 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.019773 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5wqh\" (UniqueName: \"kubernetes.io/projected/f8fbb8a7-4226-4b2f-9e95-621059e6283d-kube-api-access-d5wqh\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.019819 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-config-data\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.019842 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-log-httpd\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.019926 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-run-httpd\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.019967 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.020078 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg6rk\" (UniqueName: \"kubernetes.io/projected/65affa4a-2f58-4de2-9533-18e2d72af609-kube-api-access-jg6rk\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.020132 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-public-tls-certs\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.020193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-scripts\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.020217 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-config-data\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.020241 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.020270 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.020307 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.021626 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-log-httpd\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.021768 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-run-httpd\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.023152 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65affa4a-2f58-4de2-9533-18e2d72af609-logs\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.024879 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.025150 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-public-tls-certs\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.025181 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.026017 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.026651 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.026665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-config-data\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.026820 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-scripts\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.028009 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.036698 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-config-data\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.039875 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5wqh\" (UniqueName: \"kubernetes.io/projected/f8fbb8a7-4226-4b2f-9e95-621059e6283d-kube-api-access-d5wqh\") pod \"ceilometer-0\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.039907 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg6rk\" (UniqueName: \"kubernetes.io/projected/65affa4a-2f58-4de2-9533-18e2d72af609-kube-api-access-jg6rk\") pod \"nova-api-0\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.148496 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.160659 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.628548 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:45 crc kubenswrapper[4945]: I1206 07:19:45.711534 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:19:45 crc kubenswrapper[4945]: W1206 07:19:45.714566 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8fbb8a7_4226_4b2f_9e95_621059e6283d.slice/crio-d039eb69ff87ec7d06020cd72504dffba2a65474ffb2860d2ec9afed1e396362 WatchSource:0}: Error finding container d039eb69ff87ec7d06020cd72504dffba2a65474ffb2860d2ec9afed1e396362: Status 404 returned error can't find the container with id d039eb69ff87ec7d06020cd72504dffba2a65474ffb2860d2ec9afed1e396362 Dec 06 07:19:46 crc kubenswrapper[4945]: I1206 07:19:46.629668 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8fbb8a7-4226-4b2f-9e95-621059e6283d","Type":"ContainerStarted","Data":"d039eb69ff87ec7d06020cd72504dffba2a65474ffb2860d2ec9afed1e396362"} Dec 06 07:19:46 crc kubenswrapper[4945]: I1206 07:19:46.646583 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hssjc" podUID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerName="registry-server" containerID="cri-o://6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f" gracePeriod=2 Dec 06 07:19:46 crc kubenswrapper[4945]: I1206 07:19:46.647132 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65affa4a-2f58-4de2-9533-18e2d72af609","Type":"ContainerStarted","Data":"76064cc376884c7cfd67cd09542d1e9c457c4a281f2561b468503326a71442e8"} Dec 06 07:19:46 crc kubenswrapper[4945]: I1206 07:19:46.647267 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65affa4a-2f58-4de2-9533-18e2d72af609","Type":"ContainerStarted","Data":"615caed43e7e1d4c08c075ed058b10c254f1cb8cd2ee9f981bd1aba0970ee8e2"} Dec 06 07:19:46 crc kubenswrapper[4945]: I1206 07:19:46.647400 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65affa4a-2f58-4de2-9533-18e2d72af609","Type":"ContainerStarted","Data":"3a0873e55df6c6e47a1109007d170c06b2e4a26c3ad12bee1295e4f69f568ae6"} Dec 06 07:19:46 crc kubenswrapper[4945]: I1206 07:19:46.713011 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.712984805 podStartE2EDuration="2.712984805s" podCreationTimestamp="2025-12-06 07:19:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:46.694289179 +0000 UTC m=+1600.149150223" watchObservedRunningTime="2025-12-06 07:19:46.712984805 +0000 UTC m=+1600.167845849" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.304662 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.476059 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-catalog-content\") pod \"05954cf8-6ddf-4455-9b02-8f75b07f6272\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.476448 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzsnj\" (UniqueName: \"kubernetes.io/projected/05954cf8-6ddf-4455-9b02-8f75b07f6272-kube-api-access-bzsnj\") pod \"05954cf8-6ddf-4455-9b02-8f75b07f6272\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.476747 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-utilities\") pod \"05954cf8-6ddf-4455-9b02-8f75b07f6272\" (UID: \"05954cf8-6ddf-4455-9b02-8f75b07f6272\") " Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.477855 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-utilities" (OuterVolumeSpecName: "utilities") pod "05954cf8-6ddf-4455-9b02-8f75b07f6272" (UID: "05954cf8-6ddf-4455-9b02-8f75b07f6272"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.487238 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05954cf8-6ddf-4455-9b02-8f75b07f6272-kube-api-access-bzsnj" (OuterVolumeSpecName: "kube-api-access-bzsnj") pod "05954cf8-6ddf-4455-9b02-8f75b07f6272" (UID: "05954cf8-6ddf-4455-9b02-8f75b07f6272"). InnerVolumeSpecName "kube-api-access-bzsnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.501813 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05954cf8-6ddf-4455-9b02-8f75b07f6272" (UID: "05954cf8-6ddf-4455-9b02-8f75b07f6272"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.578954 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.579000 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzsnj\" (UniqueName: \"kubernetes.io/projected/05954cf8-6ddf-4455-9b02-8f75b07f6272-kube-api-access-bzsnj\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.579014 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05954cf8-6ddf-4455-9b02-8f75b07f6272-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.661095 4945 generic.go:334] "Generic (PLEG): container finished" podID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerID="6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f" exitCode=0 Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.661184 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hssjc" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.661190 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hssjc" event={"ID":"05954cf8-6ddf-4455-9b02-8f75b07f6272","Type":"ContainerDied","Data":"6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f"} Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.661263 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hssjc" event={"ID":"05954cf8-6ddf-4455-9b02-8f75b07f6272","Type":"ContainerDied","Data":"1cd62d9941aa0fde8450059557570108ff35355f0b45a491e20e374b2c7b1a6a"} Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.661320 4945 scope.go:117] "RemoveContainer" containerID="6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.691793 4945 scope.go:117] "RemoveContainer" containerID="46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.718470 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hssjc"] Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.731726 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hssjc"] Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.947433 4945 scope.go:117] "RemoveContainer" containerID="b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.995317 4945 scope.go:117] "RemoveContainer" containerID="6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f" Dec 06 07:19:47 crc kubenswrapper[4945]: E1206 07:19:47.995828 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f\": container with ID starting with 6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f not found: ID does not exist" containerID="6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.995939 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f"} err="failed to get container status \"6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f\": rpc error: code = NotFound desc = could not find container \"6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f\": container with ID starting with 6e6cc66a429b474505375fd1ba79e2b0b7fb378676079724dcc6035ea88ebd8f not found: ID does not exist" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.996030 4945 scope.go:117] "RemoveContainer" containerID="46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c" Dec 06 07:19:47 crc kubenswrapper[4945]: E1206 07:19:47.996632 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c\": container with ID starting with 46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c not found: ID does not exist" containerID="46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.996675 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c"} err="failed to get container status \"46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c\": rpc error: code = NotFound desc = could not find container \"46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c\": container with ID starting with 46ed0acdfb03695c8e1096c03a5f61e98e3be682c329c73582187bbd8e9c3b5c not found: ID does not exist" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.996706 4945 scope.go:117] "RemoveContainer" containerID="b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1" Dec 06 07:19:47 crc kubenswrapper[4945]: E1206 07:19:47.997025 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1\": container with ID starting with b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1 not found: ID does not exist" containerID="b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.997121 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1"} err="failed to get container status \"b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1\": rpc error: code = NotFound desc = could not find container \"b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1\": container with ID starting with b47d9c775adaac500b1c05064ddbde1cd8cc15f9c8956331bbf069bb1c94e8f1 not found: ID does not exist" Dec 06 07:19:47 crc kubenswrapper[4945]: I1206 07:19:47.998601 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:48 crc kubenswrapper[4945]: I1206 07:19:48.021124 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:48 crc kubenswrapper[4945]: I1206 07:19:48.273486 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:19:48 crc kubenswrapper[4945]: I1206 07:19:48.365584 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-fhcms"] Dec 06 07:19:48 crc kubenswrapper[4945]: I1206 07:19:48.368455 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" podUID="445042de-1881-4990-bf51-d77ec753f986" containerName="dnsmasq-dns" containerID="cri-o://8dd28a205c7b31da8ae18ba488911ee5eb0dc3e4a56c2a8ca3c0445a6e99a115" gracePeriod=10 Dec 06 07:19:48 crc kubenswrapper[4945]: I1206 07:19:48.601769 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" podUID="445042de-1881-4990-bf51-d77ec753f986" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.187:5353: connect: connection refused" Dec 06 07:19:48 crc kubenswrapper[4945]: I1206 07:19:48.683732 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8fbb8a7-4226-4b2f-9e95-621059e6283d","Type":"ContainerStarted","Data":"18675bd130adeb9a967975c3bf8070d18d7c2f22354b8443abed915b5b687156"} Dec 06 07:19:48 crc kubenswrapper[4945]: I1206 07:19:48.692166 4945 generic.go:334] "Generic (PLEG): container finished" podID="445042de-1881-4990-bf51-d77ec753f986" containerID="8dd28a205c7b31da8ae18ba488911ee5eb0dc3e4a56c2a8ca3c0445a6e99a115" exitCode=0 Dec 06 07:19:48 crc kubenswrapper[4945]: I1206 07:19:48.692238 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" event={"ID":"445042de-1881-4990-bf51-d77ec753f986","Type":"ContainerDied","Data":"8dd28a205c7b31da8ae18ba488911ee5eb0dc3e4a56c2a8ca3c0445a6e99a115"} Dec 06 07:19:48 crc kubenswrapper[4945]: I1206 07:19:48.750556 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:19:48 crc kubenswrapper[4945]: I1206 07:19:48.979464 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05954cf8-6ddf-4455-9b02-8f75b07f6272" path="/var/lib/kubelet/pods/05954cf8-6ddf-4455-9b02-8f75b07f6272/volumes" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.038089 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-g7hg8"] Dec 06 07:19:49 crc kubenswrapper[4945]: E1206 07:19:49.038621 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerName="registry-server" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.038637 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerName="registry-server" Dec 06 07:19:49 crc kubenswrapper[4945]: E1206 07:19:49.038655 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerName="extract-utilities" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.038662 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerName="extract-utilities" Dec 06 07:19:49 crc kubenswrapper[4945]: E1206 07:19:49.038681 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerName="extract-content" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.038687 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerName="extract-content" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.038886 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="05954cf8-6ddf-4455-9b02-8f75b07f6272" containerName="registry-server" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.039685 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.043269 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.043456 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.060861 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.089393 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-g7hg8"] Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.122471 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-config\") pod \"445042de-1881-4990-bf51-d77ec753f986\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.122563 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-swift-storage-0\") pod \"445042de-1881-4990-bf51-d77ec753f986\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.122688 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-svc\") pod \"445042de-1881-4990-bf51-d77ec753f986\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.122719 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-sb\") pod \"445042de-1881-4990-bf51-d77ec753f986\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.122864 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksmqf\" (UniqueName: \"kubernetes.io/projected/445042de-1881-4990-bf51-d77ec753f986-kube-api-access-ksmqf\") pod \"445042de-1881-4990-bf51-d77ec753f986\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.122930 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-nb\") pod \"445042de-1881-4990-bf51-d77ec753f986\" (UID: \"445042de-1881-4990-bf51-d77ec753f986\") " Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.124359 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-scripts\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.124500 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj5tn\" (UniqueName: \"kubernetes.io/projected/01a5e10b-0509-479c-abce-4e2b5dc9c59b-kube-api-access-bj5tn\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.124551 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-config-data\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.124577 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.157516 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/445042de-1881-4990-bf51-d77ec753f986-kube-api-access-ksmqf" (OuterVolumeSpecName: "kube-api-access-ksmqf") pod "445042de-1881-4990-bf51-d77ec753f986" (UID: "445042de-1881-4990-bf51-d77ec753f986"). InnerVolumeSpecName "kube-api-access-ksmqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.206469 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "445042de-1881-4990-bf51-d77ec753f986" (UID: "445042de-1881-4990-bf51-d77ec753f986"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.210820 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "445042de-1881-4990-bf51-d77ec753f986" (UID: "445042de-1881-4990-bf51-d77ec753f986"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.211049 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "445042de-1881-4990-bf51-d77ec753f986" (UID: "445042de-1881-4990-bf51-d77ec753f986"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.219900 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-config" (OuterVolumeSpecName: "config") pod "445042de-1881-4990-bf51-d77ec753f986" (UID: "445042de-1881-4990-bf51-d77ec753f986"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.222749 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "445042de-1881-4990-bf51-d77ec753f986" (UID: "445042de-1881-4990-bf51-d77ec753f986"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.226991 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-scripts\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.227101 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj5tn\" (UniqueName: \"kubernetes.io/projected/01a5e10b-0509-479c-abce-4e2b5dc9c59b-kube-api-access-bj5tn\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.227141 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-config-data\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.227165 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.227363 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksmqf\" (UniqueName: \"kubernetes.io/projected/445042de-1881-4990-bf51-d77ec753f986-kube-api-access-ksmqf\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.227384 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.227400 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.227412 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.227423 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.227434 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/445042de-1881-4990-bf51-d77ec753f986-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.231532 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.233835 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-scripts\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.234796 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-config-data\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.253225 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj5tn\" (UniqueName: \"kubernetes.io/projected/01a5e10b-0509-479c-abce-4e2b5dc9c59b-kube-api-access-bj5tn\") pod \"nova-cell1-cell-mapping-g7hg8\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.369824 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.708093 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8fbb8a7-4226-4b2f-9e95-621059e6283d","Type":"ContainerStarted","Data":"9e0f28c738c6aecbb802ee0805c8bd63678cb98319f526693685a2598a4babc5"} Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.708486 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8fbb8a7-4226-4b2f-9e95-621059e6283d","Type":"ContainerStarted","Data":"b9f8060a9b781ba6b1a59825c6f03e6c1703a2845c803ebc8a4b8e3e321741c0"} Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.711544 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" event={"ID":"445042de-1881-4990-bf51-d77ec753f986","Type":"ContainerDied","Data":"52eea5c2e18b5c0f7b7c7fa217b419de8092de32c9ebb410cf7e11e6ecd68ee8"} Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.711600 4945 scope.go:117] "RemoveContainer" containerID="8dd28a205c7b31da8ae18ba488911ee5eb0dc3e4a56c2a8ca3c0445a6e99a115" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.711610 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-fhcms" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.761397 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-fhcms"] Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.768026 4945 scope.go:117] "RemoveContainer" containerID="73f7cd0ccf609d91c0734d2092d72f5d6c6161a5f7adc0ef19b7dc82a92539f4" Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.772449 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-fhcms"] Dec 06 07:19:49 crc kubenswrapper[4945]: I1206 07:19:49.869317 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-g7hg8"] Dec 06 07:19:49 crc kubenswrapper[4945]: W1206 07:19:49.872289 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01a5e10b_0509_479c_abce_4e2b5dc9c59b.slice/crio-15998e758db51bc6c23a37afc1b266114065eedb0dad5a396b8d2a68338d5178 WatchSource:0}: Error finding container 15998e758db51bc6c23a37afc1b266114065eedb0dad5a396b8d2a68338d5178: Status 404 returned error can't find the container with id 15998e758db51bc6c23a37afc1b266114065eedb0dad5a396b8d2a68338d5178 Dec 06 07:19:50 crc kubenswrapper[4945]: I1206 07:19:50.723884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-g7hg8" event={"ID":"01a5e10b-0509-479c-abce-4e2b5dc9c59b","Type":"ContainerStarted","Data":"6d8dfeac270484f3d88b61ccf82778281ff7139a1cecec4590cc125407a09c91"} Dec 06 07:19:50 crc kubenswrapper[4945]: I1206 07:19:50.724233 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-g7hg8" event={"ID":"01a5e10b-0509-479c-abce-4e2b5dc9c59b","Type":"ContainerStarted","Data":"15998e758db51bc6c23a37afc1b266114065eedb0dad5a396b8d2a68338d5178"} Dec 06 07:19:50 crc kubenswrapper[4945]: I1206 07:19:50.747510 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-g7hg8" podStartSLOduration=1.747486764 podStartE2EDuration="1.747486764s" podCreationTimestamp="2025-12-06 07:19:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:19:50.739768695 +0000 UTC m=+1604.194629749" watchObservedRunningTime="2025-12-06 07:19:50.747486764 +0000 UTC m=+1604.202347798" Dec 06 07:19:50 crc kubenswrapper[4945]: I1206 07:19:50.986044 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="445042de-1881-4990-bf51-d77ec753f986" path="/var/lib/kubelet/pods/445042de-1881-4990-bf51-d77ec753f986/volumes" Dec 06 07:19:51 crc kubenswrapper[4945]: I1206 07:19:51.739864 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8fbb8a7-4226-4b2f-9e95-621059e6283d","Type":"ContainerStarted","Data":"adc583ba09127799a295f5c8a757b2e8cd330c17f41b39958b390affbc31d557"} Dec 06 07:19:51 crc kubenswrapper[4945]: I1206 07:19:51.740031 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 07:19:51 crc kubenswrapper[4945]: I1206 07:19:51.776715 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.974716876 podStartE2EDuration="7.776690608s" podCreationTimestamp="2025-12-06 07:19:44 +0000 UTC" firstStartedPulling="2025-12-06 07:19:45.72624983 +0000 UTC m=+1599.181110874" lastFinishedPulling="2025-12-06 07:19:50.528223562 +0000 UTC m=+1603.983084606" observedRunningTime="2025-12-06 07:19:51.758079714 +0000 UTC m=+1605.212940778" watchObservedRunningTime="2025-12-06 07:19:51.776690608 +0000 UTC m=+1605.231551652" Dec 06 07:19:55 crc kubenswrapper[4945]: I1206 07:19:55.161554 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 07:19:55 crc kubenswrapper[4945]: I1206 07:19:55.162087 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 07:19:55 crc kubenswrapper[4945]: I1206 07:19:55.799235 4945 generic.go:334] "Generic (PLEG): container finished" podID="01a5e10b-0509-479c-abce-4e2b5dc9c59b" containerID="6d8dfeac270484f3d88b61ccf82778281ff7139a1cecec4590cc125407a09c91" exitCode=0 Dec 06 07:19:55 crc kubenswrapper[4945]: I1206 07:19:55.799322 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-g7hg8" event={"ID":"01a5e10b-0509-479c-abce-4e2b5dc9c59b","Type":"ContainerDied","Data":"6d8dfeac270484f3d88b61ccf82778281ff7139a1cecec4590cc125407a09c91"} Dec 06 07:19:56 crc kubenswrapper[4945]: I1206 07:19:56.168482 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:19:56 crc kubenswrapper[4945]: I1206 07:19:56.174537 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.231963 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.291986 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-combined-ca-bundle\") pod \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.292052 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj5tn\" (UniqueName: \"kubernetes.io/projected/01a5e10b-0509-479c-abce-4e2b5dc9c59b-kube-api-access-bj5tn\") pod \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.292106 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-scripts\") pod \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.292134 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-config-data\") pod \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\" (UID: \"01a5e10b-0509-479c-abce-4e2b5dc9c59b\") " Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.298249 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-scripts" (OuterVolumeSpecName: "scripts") pod "01a5e10b-0509-479c-abce-4e2b5dc9c59b" (UID: "01a5e10b-0509-479c-abce-4e2b5dc9c59b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.299999 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01a5e10b-0509-479c-abce-4e2b5dc9c59b-kube-api-access-bj5tn" (OuterVolumeSpecName: "kube-api-access-bj5tn") pod "01a5e10b-0509-479c-abce-4e2b5dc9c59b" (UID: "01a5e10b-0509-479c-abce-4e2b5dc9c59b"). InnerVolumeSpecName "kube-api-access-bj5tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.326893 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01a5e10b-0509-479c-abce-4e2b5dc9c59b" (UID: "01a5e10b-0509-479c-abce-4e2b5dc9c59b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.331167 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-config-data" (OuterVolumeSpecName: "config-data") pod "01a5e10b-0509-479c-abce-4e2b5dc9c59b" (UID: "01a5e10b-0509-479c-abce-4e2b5dc9c59b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.394257 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.394330 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.394347 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a5e10b-0509-479c-abce-4e2b5dc9c59b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.394360 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj5tn\" (UniqueName: \"kubernetes.io/projected/01a5e10b-0509-479c-abce-4e2b5dc9c59b-kube-api-access-bj5tn\") on node \"crc\" DevicePath \"\"" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.821129 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-g7hg8" event={"ID":"01a5e10b-0509-479c-abce-4e2b5dc9c59b","Type":"ContainerDied","Data":"15998e758db51bc6c23a37afc1b266114065eedb0dad5a396b8d2a68338d5178"} Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.821160 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-g7hg8" Dec 06 07:19:57 crc kubenswrapper[4945]: I1206 07:19:57.821187 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15998e758db51bc6c23a37afc1b266114065eedb0dad5a396b8d2a68338d5178" Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.011099 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.011502 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" containerName="nova-api-log" containerID="cri-o://615caed43e7e1d4c08c075ed058b10c254f1cb8cd2ee9f981bd1aba0970ee8e2" gracePeriod=30 Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.011524 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" containerName="nova-api-api" containerID="cri-o://76064cc376884c7cfd67cd09542d1e9c457c4a281f2561b468503326a71442e8" gracePeriod=30 Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.044648 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.044941 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-log" containerID="cri-o://b1fda35a0d6bd8c3f59f17fe7e7d7d2a2d9557c3711c2cfd122374a43bd94380" gracePeriod=30 Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.045541 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-metadata" containerID="cri-o://a3a38b3d67d1d5ebdd251a90a3d62bade79296bcdc1f39dd668b59634f5e68f5" gracePeriod=30 Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.057951 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.061635 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5ade9957-3547-4e68-aab7-88256fddc367" containerName="nova-scheduler-scheduler" containerID="cri-o://8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6" gracePeriod=30 Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.833836 4945 generic.go:334] "Generic (PLEG): container finished" podID="65affa4a-2f58-4de2-9533-18e2d72af609" containerID="615caed43e7e1d4c08c075ed058b10c254f1cb8cd2ee9f981bd1aba0970ee8e2" exitCode=143 Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.833925 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65affa4a-2f58-4de2-9533-18e2d72af609","Type":"ContainerDied","Data":"615caed43e7e1d4c08c075ed058b10c254f1cb8cd2ee9f981bd1aba0970ee8e2"} Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.836201 4945 generic.go:334] "Generic (PLEG): container finished" podID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerID="b1fda35a0d6bd8c3f59f17fe7e7d7d2a2d9557c3711c2cfd122374a43bd94380" exitCode=143 Dec 06 07:19:58 crc kubenswrapper[4945]: I1206 07:19:58.836242 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46","Type":"ContainerDied","Data":"b1fda35a0d6bd8c3f59f17fe7e7d7d2a2d9557c3711c2cfd122374a43bd94380"} Dec 06 07:20:00 crc kubenswrapper[4945]: E1206 07:20:00.532187 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6 is running failed: container process not found" containerID="8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 07:20:00 crc kubenswrapper[4945]: E1206 07:20:00.533007 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6 is running failed: container process not found" containerID="8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 07:20:00 crc kubenswrapper[4945]: E1206 07:20:00.533350 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6 is running failed: container process not found" containerID="8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 07:20:00 crc kubenswrapper[4945]: E1206 07:20:00.533442 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="5ade9957-3547-4e68-aab7-88256fddc367" containerName="nova-scheduler-scheduler" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.595982 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.653045 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-combined-ca-bundle\") pod \"5ade9957-3547-4e68-aab7-88256fddc367\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.653103 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcvcp\" (UniqueName: \"kubernetes.io/projected/5ade9957-3547-4e68-aab7-88256fddc367-kube-api-access-zcvcp\") pod \"5ade9957-3547-4e68-aab7-88256fddc367\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.653388 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-config-data\") pod \"5ade9957-3547-4e68-aab7-88256fddc367\" (UID: \"5ade9957-3547-4e68-aab7-88256fddc367\") " Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.659690 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ade9957-3547-4e68-aab7-88256fddc367-kube-api-access-zcvcp" (OuterVolumeSpecName: "kube-api-access-zcvcp") pod "5ade9957-3547-4e68-aab7-88256fddc367" (UID: "5ade9957-3547-4e68-aab7-88256fddc367"). InnerVolumeSpecName "kube-api-access-zcvcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.682154 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ade9957-3547-4e68-aab7-88256fddc367" (UID: "5ade9957-3547-4e68-aab7-88256fddc367"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.684433 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-config-data" (OuterVolumeSpecName: "config-data") pod "5ade9957-3547-4e68-aab7-88256fddc367" (UID: "5ade9957-3547-4e68-aab7-88256fddc367"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.755606 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.755646 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcvcp\" (UniqueName: \"kubernetes.io/projected/5ade9957-3547-4e68-aab7-88256fddc367-kube-api-access-zcvcp\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.755661 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ade9957-3547-4e68-aab7-88256fddc367-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.856562 4945 generic.go:334] "Generic (PLEG): container finished" podID="5ade9957-3547-4e68-aab7-88256fddc367" containerID="8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6" exitCode=0 Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.856624 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.856619 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5ade9957-3547-4e68-aab7-88256fddc367","Type":"ContainerDied","Data":"8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6"} Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.856746 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5ade9957-3547-4e68-aab7-88256fddc367","Type":"ContainerDied","Data":"3ad6fe25b29a96b5212df8f9f2a163d388981ee36727e1d149e826c49b34b10b"} Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.856778 4945 scope.go:117] "RemoveContainer" containerID="8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.899009 4945 scope.go:117] "RemoveContainer" containerID="8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6" Dec 06 07:20:00 crc kubenswrapper[4945]: E1206 07:20:00.900068 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6\": container with ID starting with 8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6 not found: ID does not exist" containerID="8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.900176 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6"} err="failed to get container status \"8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6\": rpc error: code = NotFound desc = could not find container \"8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6\": container with ID starting with 8725fad8ee45722a7c97a039feb7cd4079fdc29eeea9545a7598d0d5232396f6 not found: ID does not exist" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.916299 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.966561 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.967776 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:20:00 crc kubenswrapper[4945]: E1206 07:20:00.968306 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445042de-1881-4990-bf51-d77ec753f986" containerName="dnsmasq-dns" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.968326 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="445042de-1881-4990-bf51-d77ec753f986" containerName="dnsmasq-dns" Dec 06 07:20:00 crc kubenswrapper[4945]: E1206 07:20:00.968347 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a5e10b-0509-479c-abce-4e2b5dc9c59b" containerName="nova-manage" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.968355 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a5e10b-0509-479c-abce-4e2b5dc9c59b" containerName="nova-manage" Dec 06 07:20:00 crc kubenswrapper[4945]: E1206 07:20:00.968374 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445042de-1881-4990-bf51-d77ec753f986" containerName="init" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.968380 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="445042de-1881-4990-bf51-d77ec753f986" containerName="init" Dec 06 07:20:00 crc kubenswrapper[4945]: E1206 07:20:00.968407 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ade9957-3547-4e68-aab7-88256fddc367" containerName="nova-scheduler-scheduler" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.968413 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ade9957-3547-4e68-aab7-88256fddc367" containerName="nova-scheduler-scheduler" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.968630 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ade9957-3547-4e68-aab7-88256fddc367" containerName="nova-scheduler-scheduler" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.968646 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="445042de-1881-4990-bf51-d77ec753f986" containerName="dnsmasq-dns" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.968671 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a5e10b-0509-479c-abce-4e2b5dc9c59b" containerName="nova-manage" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.969474 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.972378 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 07:20:00 crc kubenswrapper[4945]: I1206 07:20:00.982045 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.061832 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " pod="openstack/nova-scheduler-0" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.062002 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-config-data\") pod \"nova-scheduler-0\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " pod="openstack/nova-scheduler-0" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.062159 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp94d\" (UniqueName: \"kubernetes.io/projected/6ae75a6a-2656-4e65-a84d-8767a570ce8f-kube-api-access-vp94d\") pod \"nova-scheduler-0\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " pod="openstack/nova-scheduler-0" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.163828 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp94d\" (UniqueName: \"kubernetes.io/projected/6ae75a6a-2656-4e65-a84d-8767a570ce8f-kube-api-access-vp94d\") pod \"nova-scheduler-0\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " pod="openstack/nova-scheduler-0" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.163951 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " pod="openstack/nova-scheduler-0" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.164054 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-config-data\") pod \"nova-scheduler-0\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " pod="openstack/nova-scheduler-0" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.169736 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " pod="openstack/nova-scheduler-0" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.169874 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-config-data\") pod \"nova-scheduler-0\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " pod="openstack/nova-scheduler-0" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.183652 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp94d\" (UniqueName: \"kubernetes.io/projected/6ae75a6a-2656-4e65-a84d-8767a570ce8f-kube-api-access-vp94d\") pod \"nova-scheduler-0\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " pod="openstack/nova-scheduler-0" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.292531 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.524621 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:42570->10.217.0.191:8775: read: connection reset by peer" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.524627 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:42572->10.217.0.191:8775: read: connection reset by peer" Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.871933 4945 generic.go:334] "Generic (PLEG): container finished" podID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerID="a3a38b3d67d1d5ebdd251a90a3d62bade79296bcdc1f39dd668b59634f5e68f5" exitCode=0 Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.871989 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46","Type":"ContainerDied","Data":"a3a38b3d67d1d5ebdd251a90a3d62bade79296bcdc1f39dd668b59634f5e68f5"} Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.876375 4945 generic.go:334] "Generic (PLEG): container finished" podID="65affa4a-2f58-4de2-9533-18e2d72af609" containerID="76064cc376884c7cfd67cd09542d1e9c457c4a281f2561b468503326a71442e8" exitCode=0 Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.876461 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65affa4a-2f58-4de2-9533-18e2d72af609","Type":"ContainerDied","Data":"76064cc376884c7cfd67cd09542d1e9c457c4a281f2561b468503326a71442e8"} Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.910324 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:20:01 crc kubenswrapper[4945]: I1206 07:20:01.971775 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.025646 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.081887 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-internal-tls-certs\") pod \"65affa4a-2f58-4de2-9533-18e2d72af609\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.082044 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-public-tls-certs\") pod \"65affa4a-2f58-4de2-9533-18e2d72af609\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.082087 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-config-data\") pod \"65affa4a-2f58-4de2-9533-18e2d72af609\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.082132 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg6rk\" (UniqueName: \"kubernetes.io/projected/65affa4a-2f58-4de2-9533-18e2d72af609-kube-api-access-jg6rk\") pod \"65affa4a-2f58-4de2-9533-18e2d72af609\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.082183 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-combined-ca-bundle\") pod \"65affa4a-2f58-4de2-9533-18e2d72af609\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.082300 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f4gb\" (UniqueName: \"kubernetes.io/projected/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-kube-api-access-7f4gb\") pod \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.082364 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65affa4a-2f58-4de2-9533-18e2d72af609-logs\") pod \"65affa4a-2f58-4de2-9533-18e2d72af609\" (UID: \"65affa4a-2f58-4de2-9533-18e2d72af609\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.082418 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-combined-ca-bundle\") pod \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.082442 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-logs\") pod \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.082493 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-nova-metadata-tls-certs\") pod \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.082522 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-config-data\") pod \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\" (UID: \"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46\") " Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.090067 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65affa4a-2f58-4de2-9533-18e2d72af609-logs" (OuterVolumeSpecName: "logs") pod "65affa4a-2f58-4de2-9533-18e2d72af609" (UID: "65affa4a-2f58-4de2-9533-18e2d72af609"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.090261 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-logs" (OuterVolumeSpecName: "logs") pod "1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" (UID: "1ea42d1e-3f26-49d0-84a0-052e9bdb0e46"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.090646 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-kube-api-access-7f4gb" (OuterVolumeSpecName: "kube-api-access-7f4gb") pod "1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" (UID: "1ea42d1e-3f26-49d0-84a0-052e9bdb0e46"). InnerVolumeSpecName "kube-api-access-7f4gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.096703 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65affa4a-2f58-4de2-9533-18e2d72af609-kube-api-access-jg6rk" (OuterVolumeSpecName: "kube-api-access-jg6rk") pod "65affa4a-2f58-4de2-9533-18e2d72af609" (UID: "65affa4a-2f58-4de2-9533-18e2d72af609"). InnerVolumeSpecName "kube-api-access-jg6rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.160526 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" (UID: "1ea42d1e-3f26-49d0-84a0-052e9bdb0e46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.173569 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-config-data" (OuterVolumeSpecName: "config-data") pod "1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" (UID: "1ea42d1e-3f26-49d0-84a0-052e9bdb0e46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.177023 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65affa4a-2f58-4de2-9533-18e2d72af609" (UID: "65affa4a-2f58-4de2-9533-18e2d72af609"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.184125 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-config-data" (OuterVolumeSpecName: "config-data") pod "65affa4a-2f58-4de2-9533-18e2d72af609" (UID: "65affa4a-2f58-4de2-9533-18e2d72af609"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.185032 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.185067 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg6rk\" (UniqueName: \"kubernetes.io/projected/65affa4a-2f58-4de2-9533-18e2d72af609-kube-api-access-jg6rk\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.185079 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.185090 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f4gb\" (UniqueName: \"kubernetes.io/projected/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-kube-api-access-7f4gb\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.185130 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65affa4a-2f58-4de2-9533-18e2d72af609-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.185138 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.185148 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.185157 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.201378 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" (UID: "1ea42d1e-3f26-49d0-84a0-052e9bdb0e46"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.202548 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "65affa4a-2f58-4de2-9533-18e2d72af609" (UID: "65affa4a-2f58-4de2-9533-18e2d72af609"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.215113 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "65affa4a-2f58-4de2-9533-18e2d72af609" (UID: "65affa4a-2f58-4de2-9533-18e2d72af609"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.289092 4945 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.289120 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.289129 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65affa4a-2f58-4de2-9533-18e2d72af609-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.890838 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1ea42d1e-3f26-49d0-84a0-052e9bdb0e46","Type":"ContainerDied","Data":"25eac9f656416b22c95285047781d4afc7c4f7fe46abfaf0dc6618285ac0eb9c"} Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.890890 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.891289 4945 scope.go:117] "RemoveContainer" containerID="a3a38b3d67d1d5ebdd251a90a3d62bade79296bcdc1f39dd668b59634f5e68f5" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.892230 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ae75a6a-2656-4e65-a84d-8767a570ce8f","Type":"ContainerStarted","Data":"5984506921a2b100cf94ec4a7e72d52586d9d11cdd69b262d6ee89622fc5fcf4"} Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.892249 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ae75a6a-2656-4e65-a84d-8767a570ce8f","Type":"ContainerStarted","Data":"4a8573f8c922a12417cbb4b68c50abb4e468a708b2b1f280c89b6ccfbc242cc7"} Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.896170 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65affa4a-2f58-4de2-9533-18e2d72af609","Type":"ContainerDied","Data":"3a0873e55df6c6e47a1109007d170c06b2e4a26c3ad12bee1295e4f69f568ae6"} Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.896266 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.929235 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.929212558 podStartE2EDuration="2.929212558s" podCreationTimestamp="2025-12-06 07:20:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:20:02.923653917 +0000 UTC m=+1616.378514961" watchObservedRunningTime="2025-12-06 07:20:02.929212558 +0000 UTC m=+1616.384073602" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.929590 4945 scope.go:117] "RemoveContainer" containerID="b1fda35a0d6bd8c3f59f17fe7e7d7d2a2d9557c3711c2cfd122374a43bd94380" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.950875 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.967016 4945 scope.go:117] "RemoveContainer" containerID="76064cc376884c7cfd67cd09542d1e9c457c4a281f2561b468503326a71442e8" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.992354 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ade9957-3547-4e68-aab7-88256fddc367" path="/var/lib/kubelet/pods/5ade9957-3547-4e68-aab7-88256fddc367/volumes" Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.996046 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.996382 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:20:02 crc kubenswrapper[4945]: I1206 07:20:02.996502 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.007902 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:20:03 crc kubenswrapper[4945]: E1206 07:20:03.008940 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" containerName="nova-api-api" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.009108 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" containerName="nova-api-api" Dec 06 07:20:03 crc kubenswrapper[4945]: E1206 07:20:03.009195 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-log" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.009263 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-log" Dec 06 07:20:03 crc kubenswrapper[4945]: E1206 07:20:03.009395 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-metadata" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.009474 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-metadata" Dec 06 07:20:03 crc kubenswrapper[4945]: E1206 07:20:03.009557 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" containerName="nova-api-log" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.009631 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" containerName="nova-api-log" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.010000 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-metadata" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.010106 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" containerName="nova-metadata-log" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.010203 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" containerName="nova-api-api" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.010306 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" containerName="nova-api-log" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.012542 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.015864 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.016201 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.016565 4945 scope.go:117] "RemoveContainer" containerID="615caed43e7e1d4c08c075ed058b10c254f1cb8cd2ee9f981bd1aba0970ee8e2" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.017864 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.020073 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.029506 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.030123 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.030885 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.030961 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.044467 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.110914 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.111136 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.111169 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt4bx\" (UniqueName: \"kubernetes.io/projected/efad5b26-d064-4d06-8cc8-361d094dfa9c-kube-api-access-rt4bx\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.111200 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-config-data\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.111246 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.111330 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.111402 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.111452 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-config-data\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.111663 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efad5b26-d064-4d06-8cc8-361d094dfa9c-logs\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.111796 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-logs\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.111845 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xckkr\" (UniqueName: \"kubernetes.io/projected/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-kube-api-access-xckkr\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.214388 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.214474 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt4bx\" (UniqueName: \"kubernetes.io/projected/efad5b26-d064-4d06-8cc8-361d094dfa9c-kube-api-access-rt4bx\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.214511 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-config-data\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.214557 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.214601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.214657 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.214700 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-config-data\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.214761 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efad5b26-d064-4d06-8cc8-361d094dfa9c-logs\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.215871 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-logs\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.215933 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xckkr\" (UniqueName: \"kubernetes.io/projected/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-kube-api-access-xckkr\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.215984 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efad5b26-d064-4d06-8cc8-361d094dfa9c-logs\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.216026 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.216403 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-logs\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.221608 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.222224 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-config-data\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.225425 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.225853 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-config-data\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.229406 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.231889 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.234848 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.235264 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xckkr\" (UniqueName: \"kubernetes.io/projected/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-kube-api-access-xckkr\") pod \"nova-api-0\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.241733 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt4bx\" (UniqueName: \"kubernetes.io/projected/efad5b26-d064-4d06-8cc8-361d094dfa9c-kube-api-access-rt4bx\") pod \"nova-metadata-0\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.335770 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.363410 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.809162 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:20:03 crc kubenswrapper[4945]: W1206 07:20:03.818431 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefad5b26_d064_4d06_8cc8_361d094dfa9c.slice/crio-d57aa4467dcf950661689bb2640542998c0dd1f9ef3e33c8f455cbc0e88d8c6b WatchSource:0}: Error finding container d57aa4467dcf950661689bb2640542998c0dd1f9ef3e33c8f455cbc0e88d8c6b: Status 404 returned error can't find the container with id d57aa4467dcf950661689bb2640542998c0dd1f9ef3e33c8f455cbc0e88d8c6b Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.898729 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:20:03 crc kubenswrapper[4945]: I1206 07:20:03.907441 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"efad5b26-d064-4d06-8cc8-361d094dfa9c","Type":"ContainerStarted","Data":"d57aa4467dcf950661689bb2640542998c0dd1f9ef3e33c8f455cbc0e88d8c6b"} Dec 06 07:20:03 crc kubenswrapper[4945]: W1206 07:20:03.909145 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9296c3b2_7997_45a0_a7e9_bc1f45ec204e.slice/crio-0c9e7f39a36139d16bd7a520b178089e9c6e969515df2072ac953fdd82cb0235 WatchSource:0}: Error finding container 0c9e7f39a36139d16bd7a520b178089e9c6e969515df2072ac953fdd82cb0235: Status 404 returned error can't find the container with id 0c9e7f39a36139d16bd7a520b178089e9c6e969515df2072ac953fdd82cb0235 Dec 06 07:20:04 crc kubenswrapper[4945]: I1206 07:20:04.940579 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"efad5b26-d064-4d06-8cc8-361d094dfa9c","Type":"ContainerStarted","Data":"a7da8e3a50d0a0f1e3ae039655940701da64854727b798ffef58b2a42ba05b35"} Dec 06 07:20:04 crc kubenswrapper[4945]: I1206 07:20:04.940958 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"efad5b26-d064-4d06-8cc8-361d094dfa9c","Type":"ContainerStarted","Data":"2388a11738d51f9e160cafe82ed8ca7351af42891b7d72903a3459a847238f95"} Dec 06 07:20:04 crc kubenswrapper[4945]: I1206 07:20:04.947100 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9296c3b2-7997-45a0-a7e9-bc1f45ec204e","Type":"ContainerStarted","Data":"145d4313977864fb199b0a745f9e48fe66f8d8ec1a67bdd8264719366ba4588e"} Dec 06 07:20:04 crc kubenswrapper[4945]: I1206 07:20:04.947142 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9296c3b2-7997-45a0-a7e9-bc1f45ec204e","Type":"ContainerStarted","Data":"61ecac1b8319891d087db52d2c643d043fe8240ecef32333a71acb3632492838"} Dec 06 07:20:04 crc kubenswrapper[4945]: I1206 07:20:04.947151 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9296c3b2-7997-45a0-a7e9-bc1f45ec204e","Type":"ContainerStarted","Data":"0c9e7f39a36139d16bd7a520b178089e9c6e969515df2072ac953fdd82cb0235"} Dec 06 07:20:05 crc kubenswrapper[4945]: I1206 07:20:05.012467 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.0124306 podStartE2EDuration="3.0124306s" podCreationTimestamp="2025-12-06 07:20:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:20:04.970807607 +0000 UTC m=+1618.425668651" watchObservedRunningTime="2025-12-06 07:20:05.0124306 +0000 UTC m=+1618.467291644" Dec 06 07:20:05 crc kubenswrapper[4945]: I1206 07:20:05.023043 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.023019885 podStartE2EDuration="3.023019885s" podCreationTimestamp="2025-12-06 07:20:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:20:05.005871633 +0000 UTC m=+1618.460732677" watchObservedRunningTime="2025-12-06 07:20:05.023019885 +0000 UTC m=+1618.477880929" Dec 06 07:20:05 crc kubenswrapper[4945]: I1206 07:20:05.023690 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea42d1e-3f26-49d0-84a0-052e9bdb0e46" path="/var/lib/kubelet/pods/1ea42d1e-3f26-49d0-84a0-052e9bdb0e46/volumes" Dec 06 07:20:05 crc kubenswrapper[4945]: I1206 07:20:05.027695 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65affa4a-2f58-4de2-9533-18e2d72af609" path="/var/lib/kubelet/pods/65affa4a-2f58-4de2-9533-18e2d72af609/volumes" Dec 06 07:20:06 crc kubenswrapper[4945]: I1206 07:20:06.293551 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 07:20:08 crc kubenswrapper[4945]: I1206 07:20:08.335938 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 07:20:08 crc kubenswrapper[4945]: I1206 07:20:08.336554 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 07:20:08 crc kubenswrapper[4945]: I1206 07:20:08.795700 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:20:08 crc kubenswrapper[4945]: I1206 07:20:08.795788 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:20:11 crc kubenswrapper[4945]: I1206 07:20:11.292696 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 07:20:11 crc kubenswrapper[4945]: I1206 07:20:11.318973 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 07:20:12 crc kubenswrapper[4945]: I1206 07:20:12.040724 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 07:20:13 crc kubenswrapper[4945]: I1206 07:20:13.336861 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 07:20:13 crc kubenswrapper[4945]: I1206 07:20:13.337244 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 07:20:13 crc kubenswrapper[4945]: I1206 07:20:13.364711 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 07:20:13 crc kubenswrapper[4945]: I1206 07:20:13.364809 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 07:20:14 crc kubenswrapper[4945]: I1206 07:20:14.349529 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:20:14 crc kubenswrapper[4945]: I1206 07:20:14.349560 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:20:14 crc kubenswrapper[4945]: I1206 07:20:14.376612 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:20:14 crc kubenswrapper[4945]: I1206 07:20:14.376698 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:20:15 crc kubenswrapper[4945]: I1206 07:20:15.156654 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 07:20:23 crc kubenswrapper[4945]: I1206 07:20:23.410496 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 07:20:23 crc kubenswrapper[4945]: I1206 07:20:23.489337 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 07:20:23 crc kubenswrapper[4945]: I1206 07:20:23.490451 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 07:20:23 crc kubenswrapper[4945]: I1206 07:20:23.490829 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 07:20:23 crc kubenswrapper[4945]: I1206 07:20:23.493910 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 07:20:23 crc kubenswrapper[4945]: I1206 07:20:23.497211 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 07:20:23 crc kubenswrapper[4945]: I1206 07:20:23.497764 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 07:20:24 crc kubenswrapper[4945]: I1206 07:20:24.126739 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 07:20:24 crc kubenswrapper[4945]: I1206 07:20:24.131272 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 07:20:24 crc kubenswrapper[4945]: I1206 07:20:24.132957 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 07:20:38 crc kubenswrapper[4945]: I1206 07:20:38.796057 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:20:38 crc kubenswrapper[4945]: I1206 07:20:38.796639 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.166016 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.166752 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="7e90dc2b-f137-411b-87e4-b46f36d3034e" containerName="openstackclient" containerID="cri-o://ff0585a6020bd13a78632832e2192d413450d1f996c3ce1657896dcee244dbdb" gracePeriod=2 Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.189265 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.334287 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.335148 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="232baa8c-2d18-49f3-b230-362181e687df" containerName="ovn-northd" containerID="cri-o://01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" gracePeriod=30 Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.336762 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="232baa8c-2d18-49f3-b230-362181e687df" containerName="openstack-network-exporter" containerID="cri-o://db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba" gracePeriod=30 Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.660701 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.708107 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder3788-account-delete-56qf8"] Dec 06 07:20:44 crc kubenswrapper[4945]: E1206 07:20:44.737169 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e90dc2b-f137-411b-87e4-b46f36d3034e" containerName="openstackclient" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.737209 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e90dc2b-f137-411b-87e4-b46f36d3034e" containerName="openstackclient" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.740921 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e90dc2b-f137-411b-87e4-b46f36d3034e" containerName="openstackclient" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.742164 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder3788-account-delete-56qf8"] Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.742252 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder3788-account-delete-56qf8" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.742864 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.743065 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:20:44 crc kubenswrapper[4945]: E1206 07:20:44.799607 4945 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 06 07:20:44 crc kubenswrapper[4945]: E1206 07:20:44.808319 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data podName:250c85e2-e14d-4f93-822c-c9a88352eafe nodeName:}" failed. No retries permitted until 2025-12-06 07:20:45.308250132 +0000 UTC m=+1658.763111176 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data") pod "rabbitmq-cell1-server-0" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe") : configmap "rabbitmq-cell1-config-data" not found Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.804003 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glancef244-account-delete-vtwn6"] Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.809899 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancef244-account-delete-vtwn6" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.847919 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.848745 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" containerName="openstack-network-exporter" containerID="cri-o://8a0182213333baacc78f792add7c9120ea279ab7a0ddd78d4a1493c3debaba17" gracePeriod=300 Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.899912 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj76b\" (UniqueName: \"kubernetes.io/projected/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-kube-api-access-pj76b\") pod \"cinder3788-account-delete-56qf8\" (UID: \"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b\") " pod="openstack/cinder3788-account-delete-56qf8" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.899980 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-operator-scripts\") pod \"cinder3788-account-delete-56qf8\" (UID: \"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b\") " pod="openstack/cinder3788-account-delete-56qf8" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.900012 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x55dk\" (UniqueName: \"kubernetes.io/projected/c1cb89b2-7918-457d-a36a-2cec380f3854-kube-api-access-x55dk\") pod \"glancef244-account-delete-vtwn6\" (UID: \"c1cb89b2-7918-457d-a36a-2cec380f3854\") " pod="openstack/glancef244-account-delete-vtwn6" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.900086 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1cb89b2-7918-457d-a36a-2cec380f3854-operator-scripts\") pod \"glancef244-account-delete-vtwn6\" (UID: \"c1cb89b2-7918-457d-a36a-2cec380f3854\") " pod="openstack/glancef244-account-delete-vtwn6" Dec 06 07:20:44 crc kubenswrapper[4945]: I1206 07:20:44.950058 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancef244-account-delete-vtwn6"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.001764 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1cb89b2-7918-457d-a36a-2cec380f3854-operator-scripts\") pod \"glancef244-account-delete-vtwn6\" (UID: \"c1cb89b2-7918-457d-a36a-2cec380f3854\") " pod="openstack/glancef244-account-delete-vtwn6" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.001908 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj76b\" (UniqueName: \"kubernetes.io/projected/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-kube-api-access-pj76b\") pod \"cinder3788-account-delete-56qf8\" (UID: \"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b\") " pod="openstack/cinder3788-account-delete-56qf8" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.001941 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-operator-scripts\") pod \"cinder3788-account-delete-56qf8\" (UID: \"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b\") " pod="openstack/cinder3788-account-delete-56qf8" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.001961 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x55dk\" (UniqueName: \"kubernetes.io/projected/c1cb89b2-7918-457d-a36a-2cec380f3854-kube-api-access-x55dk\") pod \"glancef244-account-delete-vtwn6\" (UID: \"c1cb89b2-7918-457d-a36a-2cec380f3854\") " pod="openstack/glancef244-account-delete-vtwn6" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.011361 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-operator-scripts\") pod \"cinder3788-account-delete-56qf8\" (UID: \"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b\") " pod="openstack/cinder3788-account-delete-56qf8" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.014321 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1cb89b2-7918-457d-a36a-2cec380f3854-operator-scripts\") pod \"glancef244-account-delete-vtwn6\" (UID: \"c1cb89b2-7918-457d-a36a-2cec380f3854\") " pod="openstack/glancef244-account-delete-vtwn6" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.044896 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj76b\" (UniqueName: \"kubernetes.io/projected/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-kube-api-access-pj76b\") pod \"cinder3788-account-delete-56qf8\" (UID: \"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b\") " pod="openstack/cinder3788-account-delete-56qf8" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.059823 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x55dk\" (UniqueName: \"kubernetes.io/projected/c1cb89b2-7918-457d-a36a-2cec380f3854-kube-api-access-x55dk\") pod \"glancef244-account-delete-vtwn6\" (UID: \"c1cb89b2-7918-457d-a36a-2cec380f3854\") " pod="openstack/glancef244-account-delete-vtwn6" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.160908 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder3788-account-delete-56qf8" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.214871 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancef244-account-delete-vtwn6" Dec 06 07:20:45 crc kubenswrapper[4945]: E1206 07:20:45.324182 4945 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 06 07:20:45 crc kubenswrapper[4945]: E1206 07:20:45.324273 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data podName:250c85e2-e14d-4f93-822c-c9a88352eafe nodeName:}" failed. No retries permitted until 2025-12-06 07:20:46.324248834 +0000 UTC m=+1659.779109878 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data") pod "rabbitmq-cell1-server-0" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe") : configmap "rabbitmq-cell1-config-data" not found Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.417651 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" containerName="ovsdbserver-nb" containerID="cri-o://ca859b837b4d6d7c6f85efa965868ddba91015b75686b509128eaf6c975919c5" gracePeriod=300 Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.451418 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican4ffa-account-delete-r98d4"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.452629 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.452662 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican4ffa-account-delete-r98d4"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.452684 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-hr6hb"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.452698 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-hr6hb"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.452718 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement73b1-account-delete-l54br"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.453496 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican4ffa-account-delete-r98d4" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.453572 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement73b1-account-delete-l54br"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.453659 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement73b1-account-delete-l54br" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.454301 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="b51cdd1a-131a-4f4c-8521-93890fd48722" containerName="openstack-network-exporter" containerID="cri-o://dcc267ea7f865def7c69c6af3584068ad5567448a3b0f26d54246303ab0aeed6" gracePeriod=300 Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.479884 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.533678 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e3c48ac-c191-4478-877e-9600b6f63929-operator-scripts\") pod \"barbican4ffa-account-delete-r98d4\" (UID: \"5e3c48ac-c191-4478-877e-9600b6f63929\") " pod="openstack/barbican4ffa-account-delete-r98d4" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.533777 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvf82\" (UniqueName: \"kubernetes.io/projected/95dc303d-9736-4c29-a1f4-f8b362179eeb-kube-api-access-tvf82\") pod \"placement73b1-account-delete-l54br\" (UID: \"95dc303d-9736-4c29-a1f4-f8b362179eeb\") " pod="openstack/placement73b1-account-delete-l54br" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.533857 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95dc303d-9736-4c29-a1f4-f8b362179eeb-operator-scripts\") pod \"placement73b1-account-delete-l54br\" (UID: \"95dc303d-9736-4c29-a1f4-f8b362179eeb\") " pod="openstack/placement73b1-account-delete-l54br" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.533905 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wfp7\" (UniqueName: \"kubernetes.io/projected/5e3c48ac-c191-4478-877e-9600b6f63929-kube-api-access-5wfp7\") pod \"barbican4ffa-account-delete-r98d4\" (UID: \"5e3c48ac-c191-4478-877e-9600b6f63929\") " pod="openstack/barbican4ffa-account-delete-r98d4" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.538065 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-g5p9c"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.557718 4945 generic.go:334] "Generic (PLEG): container finished" podID="232baa8c-2d18-49f3-b230-362181e687df" containerID="db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba" exitCode=2 Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.557827 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"232baa8c-2d18-49f3-b230-362181e687df","Type":"ContainerDied","Data":"db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba"} Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.606442 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-g5p9c"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.621192 4945 generic.go:334] "Generic (PLEG): container finished" podID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" containerID="8a0182213333baacc78f792add7c9120ea279ab7a0ddd78d4a1493c3debaba17" exitCode=2 Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.621245 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d5496b71-8389-4d16-97c0-5752a2f1ef2a","Type":"ContainerDied","Data":"8a0182213333baacc78f792add7c9120ea279ab7a0ddd78d4a1493c3debaba17"} Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.636059 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e3c48ac-c191-4478-877e-9600b6f63929-operator-scripts\") pod \"barbican4ffa-account-delete-r98d4\" (UID: \"5e3c48ac-c191-4478-877e-9600b6f63929\") " pod="openstack/barbican4ffa-account-delete-r98d4" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.636149 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvf82\" (UniqueName: \"kubernetes.io/projected/95dc303d-9736-4c29-a1f4-f8b362179eeb-kube-api-access-tvf82\") pod \"placement73b1-account-delete-l54br\" (UID: \"95dc303d-9736-4c29-a1f4-f8b362179eeb\") " pod="openstack/placement73b1-account-delete-l54br" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.636206 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95dc303d-9736-4c29-a1f4-f8b362179eeb-operator-scripts\") pod \"placement73b1-account-delete-l54br\" (UID: \"95dc303d-9736-4c29-a1f4-f8b362179eeb\") " pod="openstack/placement73b1-account-delete-l54br" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.636239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wfp7\" (UniqueName: \"kubernetes.io/projected/5e3c48ac-c191-4478-877e-9600b6f63929-kube-api-access-5wfp7\") pod \"barbican4ffa-account-delete-r98d4\" (UID: \"5e3c48ac-c191-4478-877e-9600b6f63929\") " pod="openstack/barbican4ffa-account-delete-r98d4" Dec 06 07:20:45 crc kubenswrapper[4945]: E1206 07:20:45.638336 4945 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 06 07:20:45 crc kubenswrapper[4945]: E1206 07:20:45.638412 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data podName:d71d0084-b4bc-4be5-971b-eced6e6d66ea nodeName:}" failed. No retries permitted until 2025-12-06 07:20:46.1383857 +0000 UTC m=+1659.593246804 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data") pod "rabbitmq-server-0" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea") : configmap "rabbitmq-config-data" not found Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.638732 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95dc303d-9736-4c29-a1f4-f8b362179eeb-operator-scripts\") pod \"placement73b1-account-delete-l54br\" (UID: \"95dc303d-9736-4c29-a1f4-f8b362179eeb\") " pod="openstack/placement73b1-account-delete-l54br" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.639683 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e3c48ac-c191-4478-877e-9600b6f63929-operator-scripts\") pod \"barbican4ffa-account-delete-r98d4\" (UID: \"5e3c48ac-c191-4478-877e-9600b6f63929\") " pod="openstack/barbican4ffa-account-delete-r98d4" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.665112 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron975b-account-delete-nvsg5"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.673841 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron975b-account-delete-nvsg5" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.757548 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron975b-account-delete-nvsg5"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.772427 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvf82\" (UniqueName: \"kubernetes.io/projected/95dc303d-9736-4c29-a1f4-f8b362179eeb-kube-api-access-tvf82\") pod \"placement73b1-account-delete-l54br\" (UID: \"95dc303d-9736-4c29-a1f4-f8b362179eeb\") " pod="openstack/placement73b1-account-delete-l54br" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.772431 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wfp7\" (UniqueName: \"kubernetes.io/projected/5e3c48ac-c191-4478-877e-9600b6f63929-kube-api-access-5wfp7\") pod \"barbican4ffa-account-delete-r98d4\" (UID: \"5e3c48ac-c191-4478-877e-9600b6f63929\") " pod="openstack/barbican4ffa-account-delete-r98d4" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.797488 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-cks9w"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.849883 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts\") pod \"neutron975b-account-delete-nvsg5\" (UID: \"4231a0fc-b0c8-48da-8fab-75cd817dd3ef\") " pod="openstack/neutron975b-account-delete-nvsg5" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.850034 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfj5g\" (UniqueName: \"kubernetes.io/projected/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-kube-api-access-kfj5g\") pod \"neutron975b-account-delete-nvsg5\" (UID: \"4231a0fc-b0c8-48da-8fab-75cd817dd3ef\") " pod="openstack/neutron975b-account-delete-nvsg5" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.882351 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-cks9w"] Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.902944 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement73b1-account-delete-l54br" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.937817 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican4ffa-account-delete-r98d4" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.969680 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts\") pod \"neutron975b-account-delete-nvsg5\" (UID: \"4231a0fc-b0c8-48da-8fab-75cd817dd3ef\") " pod="openstack/neutron975b-account-delete-nvsg5" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.969816 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfj5g\" (UniqueName: \"kubernetes.io/projected/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-kube-api-access-kfj5g\") pod \"neutron975b-account-delete-nvsg5\" (UID: \"4231a0fc-b0c8-48da-8fab-75cd817dd3ef\") " pod="openstack/neutron975b-account-delete-nvsg5" Dec 06 07:20:45 crc kubenswrapper[4945]: I1206 07:20:45.984575 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-hjh59"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.004169 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts\") pod \"neutron975b-account-delete-nvsg5\" (UID: \"4231a0fc-b0c8-48da-8fab-75cd817dd3ef\") " pod="openstack/neutron975b-account-delete-nvsg5" Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.053393 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="b51cdd1a-131a-4f4c-8521-93890fd48722" containerName="ovsdbserver-sb" containerID="cri-o://4441bb3a196d951e00485c2dfb22d064eba695cdb391e51c804ea9be548b968a" gracePeriod=300 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.084268 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-5m7z8"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.087767 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfj5g\" (UniqueName: \"kubernetes.io/projected/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-kube-api-access-kfj5g\") pod \"neutron975b-account-delete-nvsg5\" (UID: \"4231a0fc-b0c8-48da-8fab-75cd817dd3ef\") " pod="openstack/neutron975b-account-delete-nvsg5" Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.127154 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-mg8gj"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.172018 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-hjh59"] Dec 06 07:20:46 crc kubenswrapper[4945]: E1206 07:20:46.191274 4945 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 06 07:20:46 crc kubenswrapper[4945]: E1206 07:20:46.191382 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data podName:d71d0084-b4bc-4be5-971b-eced6e6d66ea nodeName:}" failed. No retries permitted until 2025-12-06 07:20:47.191359921 +0000 UTC m=+1660.646220975 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data") pod "rabbitmq-server-0" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea") : configmap "rabbitmq-config-data" not found Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.191752 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-5m7z8"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.209676 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-mg8gj"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.305441 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron975b-account-delete-nvsg5" Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.310290 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-2996k"] Dec 06 07:20:46 crc kubenswrapper[4945]: E1206 07:20:46.410450 4945 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 06 07:20:46 crc kubenswrapper[4945]: E1206 07:20:46.410530 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data podName:250c85e2-e14d-4f93-822c-c9a88352eafe nodeName:}" failed. No retries permitted until 2025-12-06 07:20:48.410509464 +0000 UTC m=+1661.865370498 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data") pod "rabbitmq-cell1-server-0" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe") : configmap "rabbitmq-cell1-config-data" not found Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.411209 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xc2dd"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.457460 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-lthjk"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.457707 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-lthjk" podUID="003bd22c-19f6-4bfa-8f8d-3045dfb7c841" containerName="openstack-network-exporter" containerID="cri-o://7e2f33a76160a89b2b209c529e605d89f63fe75180ef0858d4423cb32db2c20d" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.520002 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-n747t"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.520477 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" podUID="384f1bb0-ff94-4fc4-b658-7e1a250945ae" containerName="dnsmasq-dns" containerID="cri-o://cf42f6e43b6eaa14b9d60d2d2162c0cfe2ef17791091ee3c53ee8ba8cb1d8339" gracePeriod=10 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.536344 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.536935 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-server" containerID="cri-o://8f302203e6126bf00e33188ea6dd1f7e034a8cbfe7c5182693c187c7f24b819a" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537509 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="swift-recon-cron" containerID="cri-o://789049d9004e9570b8c0301ec30ded93adc7a99fef755791d3e1828734190e19" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537571 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="rsync" containerID="cri-o://ff4e063c23aaff06ab010a43b39361561a092ed784e7fd983facb5b882fd3225" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537611 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-expirer" containerID="cri-o://6088cfa6a692765f885d01e25babb5993fb4320f6715c412196029c872327c41" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537649 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-updater" containerID="cri-o://e0fbc46abb12d20db2057256ae54a1bd2c5b3df4bd87a0d9a1700ac12c2ce2c1" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537680 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-auditor" containerID="cri-o://0c80efac78458edb45ee60a60f64b441b635ee4a49351c959c569def27d0e665" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537711 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-replicator" containerID="cri-o://a8901783d5561a470b55820422085e5ac6d8ccd2649a4eec8701a774fe804faf" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537742 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-server" containerID="cri-o://04e71c5a4df960ad72d6476b9f52dc9c05815786d261a2816740f83faab2d771" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537777 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-updater" containerID="cri-o://de0927bc9e6dd18e1a79f623e51e8901bc80dec257144ca686e887a312f70a86" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537821 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-auditor" containerID="cri-o://1991abb13d702f2e5d57933ab6d10cf0d4c07b1134ee92235420a55b874e1f9e" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537857 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-replicator" containerID="cri-o://cbfe05d94d7270d286fed8b4840eccd1701ac59c60fb2b3214890ceec204a5cd" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537887 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-server" containerID="cri-o://195d75cef2ef89afdb1df93e86f8e6e38bae58c3938182ede8e3b5daa3fde195" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537917 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-reaper" containerID="cri-o://a0fcd13fd05e6559ba8dfeed754e555d93dc6f54fcefdb814974d89ffc3bac43" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537945 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-auditor" containerID="cri-o://ff9b25f9862ebd3ee6236d1ae7fa95a0a44adbb1a4d9be0d4f84005b7298f8fd" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.537980 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-replicator" containerID="cri-o://57a847dca15d392b38d22b41bbbc40b779272394f028698e226e315d75751d47" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.546907 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.547159 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" containerName="cinder-scheduler" containerID="cri-o://14269f510b2f262408ff083c20bf00b7f68b48a2e3c497d57f6d2ea6e8cc4fcb" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.547565 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" containerName="probe" containerID="cri-o://1041e9cc315c9c6b1e1b5a4a949b745e06f5a68d6531286428930d1c1eb3b390" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.557797 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-5j7ln"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.669431 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-5j7ln"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.697358 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-g7hg8"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.739659 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-g7hg8"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.760785 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-75599744c6-dmqzn"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.761066 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-75599744c6-dmqzn" podUID="569f0e1c-e58f-4e87-bde3-e7be4659639a" containerName="placement-log" containerID="cri-o://4c7e4b215beb1d463e0865db5d67c30bce7decd535274ced24856b5651616753" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.761601 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-75599744c6-dmqzn" podUID="569f0e1c-e58f-4e87-bde3-e7be4659639a" containerName="placement-api" containerID="cri-o://54e516865e0c2bf291e50817a1f0a0df2b67a1322700ab4cf577d2385c08bcb2" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.780799 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.781051 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" containerName="glance-log" containerID="cri-o://31c19a7223a90ce2817b04d80780886488d735a425cbc88ccf01eeac5899bab6" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.781418 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" containerName="glance-httpd" containerID="cri-o://f1978a8cc52c8c7435676004e144a0b8373ccf11fd0f8cada610a31d96b4bf20" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.815029 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d5496b71-8389-4d16-97c0-5752a2f1ef2a/ovsdbserver-nb/0.log" Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.815097 4945 generic.go:334] "Generic (PLEG): container finished" podID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" containerID="ca859b837b4d6d7c6f85efa965868ddba91015b75686b509128eaf6c975919c5" exitCode=143 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.815215 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d5496b71-8389-4d16-97c0-5752a2f1ef2a","Type":"ContainerDied","Data":"ca859b837b4d6d7c6f85efa965868ddba91015b75686b509128eaf6c975919c5"} Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.863353 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.863749 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerName="cinder-api-log" containerID="cri-o://c3b041158b4f4c593ae33438ee22c67b72a78b4734276f3449241b8f0221a4dd" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.864323 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerName="cinder-api" containerID="cri-o://b957bb81cfc39bf3a8747b1e1804ba1ec5dcb33d0889082d694cf4ffd810ef17" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.872024 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.872256 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" containerName="glance-log" containerID="cri-o://41457071a0e5ace0cbbf96b7adf5ba3e23ee9463d1013d0f9d4b4f9916e1e917" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.872668 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" containerName="glance-httpd" containerID="cri-o://f75d79cb35ec6aeed2fbb5df193689624fa68ec6f36d6f07600634a6a1eb3d4f" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.878244 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b51cdd1a-131a-4f4c-8521-93890fd48722/ovsdbserver-sb/0.log" Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.878507 4945 generic.go:334] "Generic (PLEG): container finished" podID="b51cdd1a-131a-4f4c-8521-93890fd48722" containerID="dcc267ea7f865def7c69c6af3584068ad5567448a3b0f26d54246303ab0aeed6" exitCode=2 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.878522 4945 generic.go:334] "Generic (PLEG): container finished" podID="b51cdd1a-131a-4f4c-8521-93890fd48722" containerID="4441bb3a196d951e00485c2dfb22d064eba695cdb391e51c804ea9be548b968a" exitCode=143 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.878595 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b51cdd1a-131a-4f4c-8521-93890fd48722","Type":"ContainerDied","Data":"dcc267ea7f865def7c69c6af3584068ad5567448a3b0f26d54246303ab0aeed6"} Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.878623 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b51cdd1a-131a-4f4c-8521-93890fd48722","Type":"ContainerDied","Data":"4441bb3a196d951e00485c2dfb22d064eba695cdb391e51c804ea9be548b968a"} Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.888082 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0216c-account-delete-bx5zs"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.889349 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0216c-account-delete-bx5zs" Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.912249 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0216c-account-delete-bx5zs"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.931637 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p6j5\" (UniqueName: \"kubernetes.io/projected/a6fbf826-c8e6-4231-b479-90fe7de5a175-kube-api-access-4p6j5\") pod \"novacell0216c-account-delete-bx5zs\" (UID: \"a6fbf826-c8e6-4231-b479-90fe7de5a175\") " pod="openstack/novacell0216c-account-delete-bx5zs" Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.931788 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6fbf826-c8e6-4231-b479-90fe7de5a175-operator-scripts\") pod \"novacell0216c-account-delete-bx5zs\" (UID: \"a6fbf826-c8e6-4231-b479-90fe7de5a175\") " pod="openstack/novacell0216c-account-delete-bx5zs" Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.939946 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7b595c49d9-jngxp"] Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.940221 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7b595c49d9-jngxp" podUID="85ab2778-4811-4700-bca0-d55c9282699f" containerName="neutron-api" containerID="cri-o://915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.940651 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7b595c49d9-jngxp" podUID="85ab2778-4811-4700-bca0-d55c9282699f" containerName="neutron-httpd" containerID="cri-o://9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6" gracePeriod=30 Dec 06 07:20:46 crc kubenswrapper[4945]: I1206 07:20:46.951941 4945 generic.go:334] "Generic (PLEG): container finished" podID="7e90dc2b-f137-411b-87e4-b46f36d3034e" containerID="ff0585a6020bd13a78632832e2192d413450d1f996c3ce1657896dcee244dbdb" exitCode=137 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.033412 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p6j5\" (UniqueName: \"kubernetes.io/projected/a6fbf826-c8e6-4231-b479-90fe7de5a175-kube-api-access-4p6j5\") pod \"novacell0216c-account-delete-bx5zs\" (UID: \"a6fbf826-c8e6-4231-b479-90fe7de5a175\") " pod="openstack/novacell0216c-account-delete-bx5zs" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.033552 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6fbf826-c8e6-4231-b479-90fe7de5a175-operator-scripts\") pod \"novacell0216c-account-delete-bx5zs\" (UID: \"a6fbf826-c8e6-4231-b479-90fe7de5a175\") " pod="openstack/novacell0216c-account-delete-bx5zs" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.034735 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6fbf826-c8e6-4231-b479-90fe7de5a175-operator-scripts\") pod \"novacell0216c-account-delete-bx5zs\" (UID: \"a6fbf826-c8e6-4231-b479-90fe7de5a175\") " pod="openstack/novacell0216c-account-delete-bx5zs" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.088488 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p6j5\" (UniqueName: \"kubernetes.io/projected/a6fbf826-c8e6-4231-b479-90fe7de5a175-kube-api-access-4p6j5\") pod \"novacell0216c-account-delete-bx5zs\" (UID: \"a6fbf826-c8e6-4231-b479-90fe7de5a175\") " pod="openstack/novacell0216c-account-delete-bx5zs" Dec 06 07:20:47 crc kubenswrapper[4945]: E1206 07:20:47.245065 4945 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 06 07:20:47 crc kubenswrapper[4945]: E1206 07:20:47.245461 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data podName:d71d0084-b4bc-4be5-971b-eced6e6d66ea nodeName:}" failed. No retries permitted until 2025-12-06 07:20:49.245434242 +0000 UTC m=+1662.700295286 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data") pod "rabbitmq-server-0" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea") : configmap "rabbitmq-config-data" not found Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.369531 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01a5e10b-0509-479c-abce-4e2b5dc9c59b" path="/var/lib/kubelet/pods/01a5e10b-0509-479c-abce-4e2b5dc9c59b/volumes" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.372410 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3603082b-401e-484c-88a3-90c8f67f5824" path="/var/lib/kubelet/pods/3603082b-401e-484c-88a3-90c8f67f5824/volumes" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.373052 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53b7b86a-1495-4527-ad40-51ae80397771" path="/var/lib/kubelet/pods/53b7b86a-1495-4527-ad40-51ae80397771/volumes" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.373806 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa503636-fff0-4195-9364-91c2f03e5c9b" path="/var/lib/kubelet/pods/aa503636-fff0-4195-9364-91c2f03e5c9b/volumes" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.374666 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abd7c8df-1c06-4098-acc5-8ac65805e879" path="/var/lib/kubelet/pods/abd7c8df-1c06-4098-acc5-8ac65805e879/volumes" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.375835 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af445462-e7c2-481c-a777-85e8566d7b71" path="/var/lib/kubelet/pods/af445462-e7c2-481c-a777-85e8566d7b71/volumes" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.376641 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5afc31d-67d1-4faf-998c-97b660f3d510" path="/var/lib/kubelet/pods/b5afc31d-67d1-4faf-998c-97b660f3d510/volumes" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.377194 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c24e6d0b-d002-4d0c-8f6d-bb94de58fb90" path="/var/lib/kubelet/pods/c24e6d0b-d002-4d0c-8f6d-bb94de58fb90/volumes" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.378446 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.378482 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-687fd77bc7-9rms8"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.378495 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.378507 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.378519 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f9668d888-8gtj6"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.378530 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7db96d494-qst22"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.378543 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapibadc-account-delete-4n95m"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379728 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapibadc-account-delete-4n95m"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379743 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379755 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-0136-account-create-update-4q6m4"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379763 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379774 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-f5zbb"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379786 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-f5zbb"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379799 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-0136-account-create-update-4q6m4"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379809 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379819 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kd5fs"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379828 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kd5fs"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379837 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379847 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zljkx"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379856 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379816 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-log" containerID="cri-o://2388a11738d51f9e160cafe82ed8ca7351af42891b7d72903a3459a847238f95" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.380036 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="72607de1-4f23-441e-b722-3844b2eb6d15" containerName="nova-cell0-conductor-conductor" containerID="cri-o://fd592672f76730e847191d2934d9c9cfb67d21204f6a000e593a960e8df07a85" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.380195 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapibadc-account-delete-4n95m" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.380602 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-metadata" containerID="cri-o://a7da8e3a50d0a0f1e3ae039655940701da64854727b798ffef58b2a42ba05b35" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.379868 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zljkx"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.380751 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.380778 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder3788-account-delete-56qf8"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.380790 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancef244-account-delete-vtwn6"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.380946 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerName="nova-api-log" containerID="cri-o://61ecac1b8319891d087db52d2c643d043fe8240ecef32333a71acb3632492838" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.381074 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f9668d888-8gtj6" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerName="barbican-api-log" containerID="cri-o://1c796787c79166cd1f6bdca3eb60a502cd2b2fc6bd58628b9bff08d4c77ca7c4" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.381260 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3a31495b-7caf-4037-87b3-c0d7f29bf582" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://89f2192e8e4d5ac29a2fb860d1dd70435b3661183bcc6f14f8a20a90fab12155" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.381640 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerName="nova-api-api" containerID="cri-o://145d4313977864fb199b0a745f9e48fe66f8d8ec1a67bdd8264719366ba4588e" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.381747 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f9668d888-8gtj6" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerName="barbican-api" containerID="cri-o://afba2827fdc58ee118d91cef305ad6d598b1747d9ddf4810bfde53456fdf7417" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.381997 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-687fd77bc7-9rms8" podUID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" containerName="barbican-worker-log" containerID="cri-o://2326a98863fbfe1d3fcf0bf3e1db275908b04d5165d45b924f0b985263d8c616" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.382079 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="22eb2234-b14b-474c-b1bf-cf726acb064b" containerName="nova-cell1-conductor-conductor" containerID="cri-o://51e0e29bf34bea465946d1ff2713e5af788a26b0c2039d1356719fdaad5afb73" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.382117 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-687fd77bc7-9rms8" podUID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" containerName="barbican-worker" containerID="cri-o://8f06adda9c4f122a20f297ca0d7ccd2c0349a987c76bc355daf2b4ea17945592" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.382197 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6ae75a6a-2656-4e65-a84d-8767a570ce8f" containerName="nova-scheduler-scheduler" containerID="cri-o://5984506921a2b100cf94ec4a7e72d52586d9d11cdd69b262d6ee89622fc5fcf4" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.382305 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7db96d494-qst22" podUID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" containerName="barbican-keystone-listener-log" containerID="cri-o://61c8d1dfdac1f1aac44c2ab0b8448e2373849c62730037840a5d71971f8623dc" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.382386 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7db96d494-qst22" podUID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" containerName="barbican-keystone-listener" containerID="cri-o://c997070637d077fb0bd2b5c6e603ae66c9789fb87271122eac47abaf2b27318e" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.463130 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="250c85e2-e14d-4f93-822c-c9a88352eafe" containerName="rabbitmq" containerID="cri-o://5a233bae071c35e0cff40c7e5f127d5016758c87b7fe65dcf4b53dce1d196ba5" gracePeriod=604800 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.493438 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0216c-account-delete-bx5zs" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.502049 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovs-vswitchd" containerID="cri-o://59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" gracePeriod=29 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.506108 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.535220 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b51cdd1a-131a-4f4c-8521-93890fd48722/ovsdbserver-sb/0.log" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.536190 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.541084 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d5496b71-8389-4d16-97c0-5752a2f1ef2a/ovsdbserver-nb/0.log" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.541175 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.554544 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggwb9\" (UniqueName: \"kubernetes.io/projected/83eee8a2-f458-4193-8023-ca1228dc92a3-kube-api-access-ggwb9\") pod \"novaapibadc-account-delete-4n95m\" (UID: \"83eee8a2-f458-4193-8023-ca1228dc92a3\") " pod="openstack/novaapibadc-account-delete-4n95m" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.554933 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83eee8a2-f458-4193-8023-ca1228dc92a3-operator-scripts\") pod \"novaapibadc-account-delete-4n95m\" (UID: \"83eee8a2-f458-4193-8023-ca1228dc92a3\") " pod="openstack/novaapibadc-account-delete-4n95m" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.655902 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdbserver-nb-tls-certs\") pod \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.655984 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-combined-ca-bundle\") pod \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.656022 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-scripts\") pod \"b51cdd1a-131a-4f4c-8521-93890fd48722\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.656085 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqbgl\" (UniqueName: \"kubernetes.io/projected/d5496b71-8389-4d16-97c0-5752a2f1ef2a-kube-api-access-xqbgl\") pod \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.656136 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-combined-ca-bundle\") pod \"7e90dc2b-f137-411b-87e4-b46f36d3034e\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.656265 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdb-rundir\") pod \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657028 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-scripts" (OuterVolumeSpecName: "scripts") pod "b51cdd1a-131a-4f4c-8521-93890fd48722" (UID: "b51cdd1a-131a-4f4c-8521-93890fd48722"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657137 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-scripts" (OuterVolumeSpecName: "scripts") pod "d5496b71-8389-4d16-97c0-5752a2f1ef2a" (UID: "d5496b71-8389-4d16-97c0-5752a2f1ef2a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657216 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "d5496b71-8389-4d16-97c0-5752a2f1ef2a" (UID: "d5496b71-8389-4d16-97c0-5752a2f1ef2a"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.656316 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-scripts\") pod \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657347 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657370 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjvvw\" (UniqueName: \"kubernetes.io/projected/7e90dc2b-f137-411b-87e4-b46f36d3034e-kube-api-access-hjvvw\") pod \"7e90dc2b-f137-411b-87e4-b46f36d3034e\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657390 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-config\") pod \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657405 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-config\") pod \"b51cdd1a-131a-4f4c-8521-93890fd48722\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657430 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-metrics-certs-tls-certs\") pod \"b51cdd1a-131a-4f4c-8521-93890fd48722\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657448 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdbserver-sb-tls-certs\") pod \"b51cdd1a-131a-4f4c-8521-93890fd48722\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657470 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-metrics-certs-tls-certs\") pod \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\" (UID: \"d5496b71-8389-4d16-97c0-5752a2f1ef2a\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657510 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config\") pod \"7e90dc2b-f137-411b-87e4-b46f36d3034e\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657534 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbjr4\" (UniqueName: \"kubernetes.io/projected/b51cdd1a-131a-4f4c-8521-93890fd48722-kube-api-access-nbjr4\") pod \"b51cdd1a-131a-4f4c-8521-93890fd48722\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657573 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdb-rundir\") pod \"b51cdd1a-131a-4f4c-8521-93890fd48722\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657602 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config-secret\") pod \"7e90dc2b-f137-411b-87e4-b46f36d3034e\" (UID: \"7e90dc2b-f137-411b-87e4-b46f36d3034e\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.657630 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-combined-ca-bundle\") pod \"b51cdd1a-131a-4f4c-8521-93890fd48722\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.658122 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"b51cdd1a-131a-4f4c-8521-93890fd48722\" (UID: \"b51cdd1a-131a-4f4c-8521-93890fd48722\") " Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.658202 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-config" (OuterVolumeSpecName: "config") pod "d5496b71-8389-4d16-97c0-5752a2f1ef2a" (UID: "d5496b71-8389-4d16-97c0-5752a2f1ef2a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.658647 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "b51cdd1a-131a-4f4c-8521-93890fd48722" (UID: "b51cdd1a-131a-4f4c-8521-93890fd48722"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.658972 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83eee8a2-f458-4193-8023-ca1228dc92a3-operator-scripts\") pod \"novaapibadc-account-delete-4n95m\" (UID: \"83eee8a2-f458-4193-8023-ca1228dc92a3\") " pod="openstack/novaapibadc-account-delete-4n95m" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.659195 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggwb9\" (UniqueName: \"kubernetes.io/projected/83eee8a2-f458-4193-8023-ca1228dc92a3-kube-api-access-ggwb9\") pod \"novaapibadc-account-delete-4n95m\" (UID: \"83eee8a2-f458-4193-8023-ca1228dc92a3\") " pod="openstack/novaapibadc-account-delete-4n95m" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.659592 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.659617 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.659632 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5496b71-8389-4d16-97c0-5752a2f1ef2a-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.659643 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.659656 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.661059 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83eee8a2-f458-4193-8023-ca1228dc92a3-operator-scripts\") pod \"novaapibadc-account-delete-4n95m\" (UID: \"83eee8a2-f458-4193-8023-ca1228dc92a3\") " pod="openstack/novaapibadc-account-delete-4n95m" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.662045 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-config" (OuterVolumeSpecName: "config") pod "b51cdd1a-131a-4f4c-8521-93890fd48722" (UID: "b51cdd1a-131a-4f4c-8521-93890fd48722"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.762470 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b51cdd1a-131a-4f4c-8521-93890fd48722-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.834439 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6ccd85df7-25dfh"] Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.834833 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6ccd85df7-25dfh" podUID="c64f8a71-9ee5-4145-b635-bd282dabd781" containerName="proxy-server" containerID="cri-o://29a489e1a3fc4b892c014063d26ad8a5a7da41074019dcb533e94027b8e256f2" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.834786 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6ccd85df7-25dfh" podUID="c64f8a71-9ee5-4145-b635-bd282dabd781" containerName="proxy-httpd" containerID="cri-o://6afab1fa21766e8cbfdedb4c1868e051427546f3e9b764dfcadfaac9ac9f3aa3" gracePeriod=30 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.963049 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.963112 4945 scope.go:117] "RemoveContainer" containerID="ff0585a6020bd13a78632832e2192d413450d1f996c3ce1657896dcee244dbdb" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.977898 4945 generic.go:334] "Generic (PLEG): container finished" podID="85ab2778-4811-4700-bca0-d55c9282699f" containerID="9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.977961 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b595c49d9-jngxp" event={"ID":"85ab2778-4811-4700-bca0-d55c9282699f","Type":"ContainerDied","Data":"9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.979816 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancef244-account-delete-vtwn6" event={"ID":"c1cb89b2-7918-457d-a36a-2cec380f3854","Type":"ContainerStarted","Data":"6a3b2affe563ac0f25fee01855b43264336522f34295ffc38cae32aa2dd70751"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.983262 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b51cdd1a-131a-4f4c-8521-93890fd48722/ovsdbserver-sb/0.log" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.983395 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b51cdd1a-131a-4f4c-8521-93890fd48722","Type":"ContainerDied","Data":"822d32044760edff7f3133ad6da795c622a90d0cd824f06234ebbfefb18b49e0"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.983438 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.985196 4945 generic.go:334] "Generic (PLEG): container finished" podID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" containerID="41457071a0e5ace0cbbf96b7adf5ba3e23ee9463d1013d0f9d4b4f9916e1e917" exitCode=143 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.985271 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ba3e057-d321-48ca-b7a0-8d791cbb7169","Type":"ContainerDied","Data":"41457071a0e5ace0cbbf96b7adf5ba3e23ee9463d1013d0f9d4b4f9916e1e917"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.987261 4945 generic.go:334] "Generic (PLEG): container finished" podID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" containerID="31c19a7223a90ce2817b04d80780886488d735a425cbc88ccf01eeac5899bab6" exitCode=143 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.987311 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf0fa4a-4a9c-4644-9fe3-413500c59a98","Type":"ContainerDied","Data":"31c19a7223a90ce2817b04d80780886488d735a425cbc88ccf01eeac5899bab6"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.988505 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lthjk_003bd22c-19f6-4bfa-8f8d-3045dfb7c841/openstack-network-exporter/0.log" Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.988538 4945 generic.go:334] "Generic (PLEG): container finished" podID="003bd22c-19f6-4bfa-8f8d-3045dfb7c841" containerID="7e2f33a76160a89b2b209c529e605d89f63fe75180ef0858d4423cb32db2c20d" exitCode=2 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.988577 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lthjk" event={"ID":"003bd22c-19f6-4bfa-8f8d-3045dfb7c841","Type":"ContainerDied","Data":"7e2f33a76160a89b2b209c529e605d89f63fe75180ef0858d4423cb32db2c20d"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.990205 4945 generic.go:334] "Generic (PLEG): container finished" podID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerID="c3b041158b4f4c593ae33438ee22c67b72a78b4734276f3449241b8f0221a4dd" exitCode=143 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.990262 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"edbffd82-c356-43a3-94a3-3549b7d9cd31","Type":"ContainerDied","Data":"c3b041158b4f4c593ae33438ee22c67b72a78b4734276f3449241b8f0221a4dd"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994176 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="ff4e063c23aaff06ab010a43b39361561a092ed784e7fd983facb5b882fd3225" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994211 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="6088cfa6a692765f885d01e25babb5993fb4320f6715c412196029c872327c41" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994220 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="e0fbc46abb12d20db2057256ae54a1bd2c5b3df4bd87a0d9a1700ac12c2ce2c1" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994227 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="0c80efac78458edb45ee60a60f64b441b635ee4a49351c959c569def27d0e665" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994233 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="a8901783d5561a470b55820422085e5ac6d8ccd2649a4eec8701a774fe804faf" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994239 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="04e71c5a4df960ad72d6476b9f52dc9c05815786d261a2816740f83faab2d771" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994246 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="de0927bc9e6dd18e1a79f623e51e8901bc80dec257144ca686e887a312f70a86" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994253 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="1991abb13d702f2e5d57933ab6d10cf0d4c07b1134ee92235420a55b874e1f9e" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994260 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="cbfe05d94d7270d286fed8b4840eccd1701ac59c60fb2b3214890ceec204a5cd" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994266 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="195d75cef2ef89afdb1df93e86f8e6e38bae58c3938182ede8e3b5daa3fde195" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994273 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="a0fcd13fd05e6559ba8dfeed754e555d93dc6f54fcefdb814974d89ffc3bac43" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994295 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="ff9b25f9862ebd3ee6236d1ae7fa95a0a44adbb1a4d9be0d4f84005b7298f8fd" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994303 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="57a847dca15d392b38d22b41bbbc40b779272394f028698e226e315d75751d47" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994309 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="8f302203e6126bf00e33188ea6dd1f7e034a8cbfe7c5182693c187c7f24b819a" exitCode=0 Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994354 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"ff4e063c23aaff06ab010a43b39361561a092ed784e7fd983facb5b882fd3225"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994385 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"6088cfa6a692765f885d01e25babb5993fb4320f6715c412196029c872327c41"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994396 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"e0fbc46abb12d20db2057256ae54a1bd2c5b3df4bd87a0d9a1700ac12c2ce2c1"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994405 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"0c80efac78458edb45ee60a60f64b441b635ee4a49351c959c569def27d0e665"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994414 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"a8901783d5561a470b55820422085e5ac6d8ccd2649a4eec8701a774fe804faf"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994422 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"04e71c5a4df960ad72d6476b9f52dc9c05815786d261a2816740f83faab2d771"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994430 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"de0927bc9e6dd18e1a79f623e51e8901bc80dec257144ca686e887a312f70a86"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994440 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"1991abb13d702f2e5d57933ab6d10cf0d4c07b1134ee92235420a55b874e1f9e"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994450 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"cbfe05d94d7270d286fed8b4840eccd1701ac59c60fb2b3214890ceec204a5cd"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994460 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"195d75cef2ef89afdb1df93e86f8e6e38bae58c3938182ede8e3b5daa3fde195"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994470 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"a0fcd13fd05e6559ba8dfeed754e555d93dc6f54fcefdb814974d89ffc3bac43"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994478 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"ff9b25f9862ebd3ee6236d1ae7fa95a0a44adbb1a4d9be0d4f84005b7298f8fd"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994487 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"57a847dca15d392b38d22b41bbbc40b779272394f028698e226e315d75751d47"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.994495 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"8f302203e6126bf00e33188ea6dd1f7e034a8cbfe7c5182693c187c7f24b819a"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.995553 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder3788-account-delete-56qf8" event={"ID":"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b","Type":"ContainerStarted","Data":"db5de8009af333da3fb0800bc459c1663d4ea20f27eee067c31e33b99be107fc"} Dec 06 07:20:47 crc kubenswrapper[4945]: I1206 07:20:47.998733 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="3a31495b-7caf-4037-87b3-c0d7f29bf582" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.196:6080/vnc_lite.html\": dial tcp 10.217.0.196:6080: connect: connection refused" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:47.999977 4945 generic.go:334] "Generic (PLEG): container finished" podID="384f1bb0-ff94-4fc4-b658-7e1a250945ae" containerID="cf42f6e43b6eaa14b9d60d2d2162c0cfe2ef17791091ee3c53ee8ba8cb1d8339" exitCode=0 Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.000058 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" event={"ID":"384f1bb0-ff94-4fc4-b658-7e1a250945ae","Type":"ContainerDied","Data":"cf42f6e43b6eaa14b9d60d2d2162c0cfe2ef17791091ee3c53ee8ba8cb1d8339"} Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.002354 4945 generic.go:334] "Generic (PLEG): container finished" podID="569f0e1c-e58f-4e87-bde3-e7be4659639a" containerID="4c7e4b215beb1d463e0865db5d67c30bce7decd535274ced24856b5651616753" exitCode=143 Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.002433 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75599744c6-dmqzn" event={"ID":"569f0e1c-e58f-4e87-bde3-e7be4659639a","Type":"ContainerDied","Data":"4c7e4b215beb1d463e0865db5d67c30bce7decd535274ced24856b5651616753"} Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.004264 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d5496b71-8389-4d16-97c0-5752a2f1ef2a/ovsdbserver-nb/0.log" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.004339 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d5496b71-8389-4d16-97c0-5752a2f1ef2a","Type":"ContainerDied","Data":"48a0d29b488df48720b7b28465ae6845e51f0c383c1cee85f2166c39938f50df"} Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.004414 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.091369 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "d5496b71-8389-4d16-97c0-5752a2f1ef2a" (UID: "d5496b71-8389-4d16-97c0-5752a2f1ef2a"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.091485 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "b51cdd1a-131a-4f4c-8521-93890fd48722" (UID: "b51cdd1a-131a-4f4c-8521-93890fd48722"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.091586 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5496b71-8389-4d16-97c0-5752a2f1ef2a-kube-api-access-xqbgl" (OuterVolumeSpecName: "kube-api-access-xqbgl") pod "d5496b71-8389-4d16-97c0-5752a2f1ef2a" (UID: "d5496b71-8389-4d16-97c0-5752a2f1ef2a"). InnerVolumeSpecName "kube-api-access-xqbgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.091648 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b51cdd1a-131a-4f4c-8521-93890fd48722-kube-api-access-nbjr4" (OuterVolumeSpecName: "kube-api-access-nbjr4") pod "b51cdd1a-131a-4f4c-8521-93890fd48722" (UID: "b51cdd1a-131a-4f4c-8521-93890fd48722"). InnerVolumeSpecName "kube-api-access-nbjr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.092740 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggwb9\" (UniqueName: \"kubernetes.io/projected/83eee8a2-f458-4193-8023-ca1228dc92a3-kube-api-access-ggwb9\") pod \"novaapibadc-account-delete-4n95m\" (UID: \"83eee8a2-f458-4193-8023-ca1228dc92a3\") " pod="openstack/novaapibadc-account-delete-4n95m" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.093516 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e90dc2b-f137-411b-87e4-b46f36d3034e-kube-api-access-hjvvw" (OuterVolumeSpecName: "kube-api-access-hjvvw") pod "7e90dc2b-f137-411b-87e4-b46f36d3034e" (UID: "7e90dc2b-f137-411b-87e4-b46f36d3034e"). InnerVolumeSpecName "kube-api-access-hjvvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.166789 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" containerName="rabbitmq" containerID="cri-o://35755191ab63b793ba68c4e84cc61180235cdedf764a176765a67c7090121592" gracePeriod=604800 Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.173263 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapibadc-account-delete-4n95m" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.182548 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbjr4\" (UniqueName: \"kubernetes.io/projected/b51cdd1a-131a-4f4c-8521-93890fd48722-kube-api-access-nbjr4\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.182617 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.182635 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqbgl\" (UniqueName: \"kubernetes.io/projected/d5496b71-8389-4d16-97c0-5752a2f1ef2a-kube-api-access-xqbgl\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.182658 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.182747 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjvvw\" (UniqueName: \"kubernetes.io/projected/7e90dc2b-f137-411b-87e4-b46f36d3034e-kube-api-access-hjvvw\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.272074 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" podUID="384f1bb0-ff94-4fc4-b658-7e1a250945ae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.197:5353: connect: connection refused" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.319380 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.386204 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "7e90dc2b-f137-411b-87e4-b46f36d3034e" (UID: "7e90dc2b-f137-411b-87e4-b46f36d3034e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.404475 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b51cdd1a-131a-4f4c-8521-93890fd48722" (UID: "b51cdd1a-131a-4f4c-8521-93890fd48722"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: E1206 07:20:48.420885 4945 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 06 07:20:48 crc kubenswrapper[4945]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 06 07:20:48 crc kubenswrapper[4945]: + source /usr/local/bin/container-scripts/functions Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNBridge=br-int Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNRemote=tcp:localhost:6642 Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNEncapType=geneve Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNAvailabilityZones= Dec 06 07:20:48 crc kubenswrapper[4945]: ++ EnableChassisAsGateway=true Dec 06 07:20:48 crc kubenswrapper[4945]: ++ PhysicalNetworks= Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNHostName= Dec 06 07:20:48 crc kubenswrapper[4945]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 06 07:20:48 crc kubenswrapper[4945]: ++ ovs_dir=/var/lib/openvswitch Dec 06 07:20:48 crc kubenswrapper[4945]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 06 07:20:48 crc kubenswrapper[4945]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 06 07:20:48 crc kubenswrapper[4945]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 06 07:20:48 crc kubenswrapper[4945]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 06 07:20:48 crc kubenswrapper[4945]: + sleep 0.5 Dec 06 07:20:48 crc kubenswrapper[4945]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 06 07:20:48 crc kubenswrapper[4945]: + sleep 0.5 Dec 06 07:20:48 crc kubenswrapper[4945]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 06 07:20:48 crc kubenswrapper[4945]: + cleanup_ovsdb_server_semaphore Dec 06 07:20:48 crc kubenswrapper[4945]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 06 07:20:48 crc kubenswrapper[4945]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 06 07:20:48 crc kubenswrapper[4945]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-2996k" message=< Dec 06 07:20:48 crc kubenswrapper[4945]: Exiting ovsdb-server (5) ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 06 07:20:48 crc kubenswrapper[4945]: + source /usr/local/bin/container-scripts/functions Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNBridge=br-int Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNRemote=tcp:localhost:6642 Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNEncapType=geneve Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNAvailabilityZones= Dec 06 07:20:48 crc kubenswrapper[4945]: ++ EnableChassisAsGateway=true Dec 06 07:20:48 crc kubenswrapper[4945]: ++ PhysicalNetworks= Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNHostName= Dec 06 07:20:48 crc kubenswrapper[4945]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 06 07:20:48 crc kubenswrapper[4945]: ++ ovs_dir=/var/lib/openvswitch Dec 06 07:20:48 crc kubenswrapper[4945]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 06 07:20:48 crc kubenswrapper[4945]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 06 07:20:48 crc kubenswrapper[4945]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 06 07:20:48 crc kubenswrapper[4945]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 06 07:20:48 crc kubenswrapper[4945]: + sleep 0.5 Dec 06 07:20:48 crc kubenswrapper[4945]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 06 07:20:48 crc kubenswrapper[4945]: + sleep 0.5 Dec 06 07:20:48 crc kubenswrapper[4945]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 06 07:20:48 crc kubenswrapper[4945]: + cleanup_ovsdb_server_semaphore Dec 06 07:20:48 crc kubenswrapper[4945]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 06 07:20:48 crc kubenswrapper[4945]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 06 07:20:48 crc kubenswrapper[4945]: > Dec 06 07:20:48 crc kubenswrapper[4945]: E1206 07:20:48.420939 4945 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 06 07:20:48 crc kubenswrapper[4945]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 06 07:20:48 crc kubenswrapper[4945]: + source /usr/local/bin/container-scripts/functions Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNBridge=br-int Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNRemote=tcp:localhost:6642 Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNEncapType=geneve Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNAvailabilityZones= Dec 06 07:20:48 crc kubenswrapper[4945]: ++ EnableChassisAsGateway=true Dec 06 07:20:48 crc kubenswrapper[4945]: ++ PhysicalNetworks= Dec 06 07:20:48 crc kubenswrapper[4945]: ++ OVNHostName= Dec 06 07:20:48 crc kubenswrapper[4945]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 06 07:20:48 crc kubenswrapper[4945]: ++ ovs_dir=/var/lib/openvswitch Dec 06 07:20:48 crc kubenswrapper[4945]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 06 07:20:48 crc kubenswrapper[4945]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 06 07:20:48 crc kubenswrapper[4945]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 06 07:20:48 crc kubenswrapper[4945]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 06 07:20:48 crc kubenswrapper[4945]: + sleep 0.5 Dec 06 07:20:48 crc kubenswrapper[4945]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 06 07:20:48 crc kubenswrapper[4945]: + sleep 0.5 Dec 06 07:20:48 crc kubenswrapper[4945]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 06 07:20:48 crc kubenswrapper[4945]: + cleanup_ovsdb_server_semaphore Dec 06 07:20:48 crc kubenswrapper[4945]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 06 07:20:48 crc kubenswrapper[4945]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 06 07:20:48 crc kubenswrapper[4945]: > pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" containerID="cri-o://ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.420972 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" containerID="cri-o://ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" gracePeriod=28 Dec 06 07:20:48 crc kubenswrapper[4945]: E1206 07:20:48.426677 4945 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 06 07:20:48 crc kubenswrapper[4945]: E1206 07:20:48.426780 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data podName:250c85e2-e14d-4f93-822c-c9a88352eafe nodeName:}" failed. No retries permitted until 2025-12-06 07:20:52.426758125 +0000 UTC m=+1665.881619169 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data") pod "rabbitmq-cell1-server-0" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe") : configmap "rabbitmq-cell1-config-data" not found Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.432427 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.432478 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.432488 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.431488 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e90dc2b-f137-411b-87e4-b46f36d3034e" (UID: "7e90dc2b-f137-411b-87e4-b46f36d3034e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.443490 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican4ffa-account-delete-r98d4"] Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.506256 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.517331 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7e90dc2b-f137-411b-87e4-b46f36d3034e" (UID: "7e90dc2b-f137-411b-87e4-b46f36d3034e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.534650 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.534678 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.534689 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e90dc2b-f137-411b-87e4-b46f36d3034e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.602195 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5496b71-8389-4d16-97c0-5752a2f1ef2a" (UID: "d5496b71-8389-4d16-97c0-5752a2f1ef2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.677073 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.681515 4945 scope.go:117] "RemoveContainer" containerID="dcc267ea7f865def7c69c6af3584068ad5567448a3b0f26d54246303ab0aeed6" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.686801 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lthjk_003bd22c-19f6-4bfa-8f8d-3045dfb7c841/openstack-network-exporter/0.log" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.686881 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:20:48 crc kubenswrapper[4945]: E1206 07:20:48.697629 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.713462 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "d5496b71-8389-4d16-97c0-5752a2f1ef2a" (UID: "d5496b71-8389-4d16-97c0-5752a2f1ef2a"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: E1206 07:20:48.715473 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 06 07:20:48 crc kubenswrapper[4945]: E1206 07:20:48.731976 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 06 07:20:48 crc kubenswrapper[4945]: E1206 07:20:48.732072 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="232baa8c-2d18-49f3-b230-362181e687df" containerName="ovn-northd" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.747251 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" containerName="galera" containerID="cri-o://bb3773d5b61628e5effd422bfa3be33b5be1e9ba5f96b66980463e518f4da6ac" gracePeriod=29 Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.777713 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovn-rundir\") pod \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.777849 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-combined-ca-bundle\") pod \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.777955 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zllcm\" (UniqueName: \"kubernetes.io/projected/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-kube-api-access-zllcm\") pod \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.778023 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-config\") pod \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.778076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovs-rundir\") pod \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.778095 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-metrics-certs-tls-certs\") pod \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\" (UID: \"003bd22c-19f6-4bfa-8f8d-3045dfb7c841\") " Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.778548 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.782911 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-config" (OuterVolumeSpecName: "config") pod "003bd22c-19f6-4bfa-8f8d-3045dfb7c841" (UID: "003bd22c-19f6-4bfa-8f8d-3045dfb7c841"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.783070 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "003bd22c-19f6-4bfa-8f8d-3045dfb7c841" (UID: "003bd22c-19f6-4bfa-8f8d-3045dfb7c841"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.792434 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "003bd22c-19f6-4bfa-8f8d-3045dfb7c841" (UID: "003bd22c-19f6-4bfa-8f8d-3045dfb7c841"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.810382 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-kube-api-access-zllcm" (OuterVolumeSpecName: "kube-api-access-zllcm") pod "003bd22c-19f6-4bfa-8f8d-3045dfb7c841" (UID: "003bd22c-19f6-4bfa-8f8d-3045dfb7c841"). InnerVolumeSpecName "kube-api-access-zllcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.819827 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement73b1-account-delete-l54br"] Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.882481 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zllcm\" (UniqueName: \"kubernetes.io/projected/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-kube-api-access-zllcm\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.882816 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.882830 4945 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovs-rundir\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:48 crc kubenswrapper[4945]: I1206 07:20:48.882840 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.030320 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "b51cdd1a-131a-4f4c-8521-93890fd48722" (UID: "b51cdd1a-131a-4f4c-8521-93890fd48722"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.031663 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "003bd22c-19f6-4bfa-8f8d-3045dfb7c841" (UID: "003bd22c-19f6-4bfa-8f8d-3045dfb7c841"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.033362 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "d5496b71-8389-4d16-97c0-5752a2f1ef2a" (UID: "d5496b71-8389-4d16-97c0-5752a2f1ef2a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.036798 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72cf544a-1950-42b5-b244-b20339a8c702" path="/var/lib/kubelet/pods/72cf544a-1950-42b5-b244-b20339a8c702/volumes" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.050247 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e90dc2b-f137-411b-87e4-b46f36d3034e" path="/var/lib/kubelet/pods/7e90dc2b-f137-411b-87e4-b46f36d3034e/volumes" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.053579 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553" path="/var/lib/kubelet/pods/83c2ac1d-e5c9-4c4f-8ae5-15e9d7f47553/volumes" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.055151 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b017666f-1224-49c2-842c-d2e0034556e7" path="/var/lib/kubelet/pods/b017666f-1224-49c2-842c-d2e0034556e7/volumes" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.055834 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b" path="/var/lib/kubelet/pods/b8b7d3fd-fcf0-4f38-b2d1-31b883b0922b/volumes" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.092444 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.092484 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5496b71-8389-4d16-97c0-5752a2f1ef2a-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.092497 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.099616 4945 generic.go:334] "Generic (PLEG): container finished" podID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" exitCode=0 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.109488 4945 generic.go:334] "Generic (PLEG): container finished" podID="c64f8a71-9ee5-4145-b635-bd282dabd781" containerID="29a489e1a3fc4b892c014063d26ad8a5a7da41074019dcb533e94027b8e256f2" exitCode=0 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.109618 4945 generic.go:334] "Generic (PLEG): container finished" podID="c64f8a71-9ee5-4145-b635-bd282dabd781" containerID="6afab1fa21766e8cbfdedb4c1868e051427546f3e9b764dfcadfaac9ac9f3aa3" exitCode=0 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.111951 4945 generic.go:334] "Generic (PLEG): container finished" podID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" containerID="2326a98863fbfe1d3fcf0bf3e1db275908b04d5165d45b924f0b985263d8c616" exitCode=143 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.121304 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerID="1c796787c79166cd1f6bdca3eb60a502cd2b2fc6bd58628b9bff08d4c77ca7c4" exitCode=143 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.132573 4945 generic.go:334] "Generic (PLEG): container finished" podID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" containerID="1041e9cc315c9c6b1e1b5a4a949b745e06f5a68d6531286428930d1c1eb3b390" exitCode=0 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.134825 4945 generic.go:334] "Generic (PLEG): container finished" podID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerID="2388a11738d51f9e160cafe82ed8ca7351af42891b7d72903a3459a847238f95" exitCode=143 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.138223 4945 generic.go:334] "Generic (PLEG): container finished" podID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerID="61ecac1b8319891d087db52d2c643d043fe8240ecef32333a71acb3632492838" exitCode=143 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.155172 4945 generic.go:334] "Generic (PLEG): container finished" podID="3a31495b-7caf-4037-87b3-c0d7f29bf582" containerID="89f2192e8e4d5ac29a2fb860d1dd70435b3661183bcc6f14f8a20a90fab12155" exitCode=0 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.157469 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lthjk_003bd22c-19f6-4bfa-8f8d-3045dfb7c841/openstack-network-exporter/0.log" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.157593 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lthjk" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.168941 4945 generic.go:334] "Generic (PLEG): container finished" podID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" containerID="61c8d1dfdac1f1aac44c2ab0b8448e2373849c62730037840a5d71971f8623dc" exitCode=143 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.199734 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "b51cdd1a-131a-4f4c-8521-93890fd48722" (UID: "b51cdd1a-131a-4f4c-8521-93890fd48722"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.223523 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "003bd22c-19f6-4bfa-8f8d-3045dfb7c841" (UID: "003bd22c-19f6-4bfa-8f8d-3045dfb7c841"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.300266 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/003bd22c-19f6-4bfa-8f8d-3045dfb7c841-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.300584 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51cdd1a-131a-4f4c-8521-93890fd48722-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: E1206 07:20:49.300327 4945 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 06 07:20:49 crc kubenswrapper[4945]: E1206 07:20:49.300657 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data podName:d71d0084-b4bc-4be5-971b-eced6e6d66ea nodeName:}" failed. No retries permitted until 2025-12-06 07:20:53.300635274 +0000 UTC m=+1666.755496318 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data") pod "rabbitmq-server-0" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea") : configmap "rabbitmq-config-data" not found Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376778 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican4ffa-account-delete-r98d4" event={"ID":"5e3c48ac-c191-4478-877e-9600b6f63929","Type":"ContainerStarted","Data":"16290c3dc010794f4acd169c72aa06539ee1641e2027d05f2b3c23d813040254"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376821 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2996k" event={"ID":"9c092be9-37db-49bb-be6b-56ce2e0e15fd","Type":"ContainerDied","Data":"ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376840 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement73b1-account-delete-l54br" event={"ID":"95dc303d-9736-4c29-a1f4-f8b362179eeb","Type":"ContainerStarted","Data":"9f3461d0026224bb1e1e73854fad6043ef5f472047ddc7b37ea6f67d44b0fc12"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376857 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6ccd85df7-25dfh" event={"ID":"c64f8a71-9ee5-4145-b635-bd282dabd781","Type":"ContainerDied","Data":"29a489e1a3fc4b892c014063d26ad8a5a7da41074019dcb533e94027b8e256f2"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376871 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6ccd85df7-25dfh" event={"ID":"c64f8a71-9ee5-4145-b635-bd282dabd781","Type":"ContainerDied","Data":"6afab1fa21766e8cbfdedb4c1868e051427546f3e9b764dfcadfaac9ac9f3aa3"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376881 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-687fd77bc7-9rms8" event={"ID":"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac","Type":"ContainerDied","Data":"2326a98863fbfe1d3fcf0bf3e1db275908b04d5165d45b924f0b985263d8c616"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" event={"ID":"384f1bb0-ff94-4fc4-b658-7e1a250945ae","Type":"ContainerDied","Data":"c6628efe2ca405c74026daa5ade200e1299c7929321b142e8f16513a47a06e6c"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376907 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6628efe2ca405c74026daa5ade200e1299c7929321b142e8f16513a47a06e6c" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376917 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9668d888-8gtj6" event={"ID":"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975","Type":"ContainerDied","Data":"1c796787c79166cd1f6bdca3eb60a502cd2b2fc6bd58628b9bff08d4c77ca7c4"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376929 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b6d74c1f-98d7-40ac-a1e8-233604971bc9","Type":"ContainerDied","Data":"1041e9cc315c9c6b1e1b5a4a949b745e06f5a68d6531286428930d1c1eb3b390"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376942 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"efad5b26-d064-4d06-8cc8-361d094dfa9c","Type":"ContainerDied","Data":"2388a11738d51f9e160cafe82ed8ca7351af42891b7d72903a3459a847238f95"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376954 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9296c3b2-7997-45a0-a7e9-bc1f45ec204e","Type":"ContainerDied","Data":"61ecac1b8319891d087db52d2c643d043fe8240ecef32333a71acb3632492838"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376968 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a31495b-7caf-4037-87b3-c0d7f29bf582","Type":"ContainerDied","Data":"89f2192e8e4d5ac29a2fb860d1dd70435b3661183bcc6f14f8a20a90fab12155"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376981 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lthjk" event={"ID":"003bd22c-19f6-4bfa-8f8d-3045dfb7c841","Type":"ContainerDied","Data":"0327e480f210fd69e40002956bec2d16f8d2d372efb52eb02f84e6de3499153a"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.376995 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7db96d494-qst22" event={"ID":"966c9b21-65c4-4d2a-af0f-0041a9bfea08","Type":"ContainerDied","Data":"61c8d1dfdac1f1aac44c2ab0b8448e2373849c62730037840a5d71971f8623dc"} Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.404239 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron975b-account-delete-nvsg5"] Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.406949 4945 scope.go:117] "RemoveContainer" containerID="4441bb3a196d951e00485c2dfb22d064eba695cdb391e51c804ea9be548b968a" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.409919 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.470673 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0216c-account-delete-bx5zs"] Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.484334 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.491196 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.504345 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.509470 4945 scope.go:117] "RemoveContainer" containerID="8a0182213333baacc78f792add7c9120ea279ab7a0ddd78d4a1493c3debaba17" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.546886 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-lthjk"] Dec 06 07:20:49 crc kubenswrapper[4945]: W1206 07:20:49.552226 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6fbf826_c8e6_4231_b479_90fe7de5a175.slice/crio-2aceb45c5bf7d83622fb1a78f07382add9a1018ebff2cef376918ef0889ca602 WatchSource:0}: Error finding container 2aceb45c5bf7d83622fb1a78f07382add9a1018ebff2cef376918ef0889ca602: Status 404 returned error can't find the container with id 2aceb45c5bf7d83622fb1a78f07382add9a1018ebff2cef376918ef0889ca602 Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.557290 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-lthjk"] Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.558138 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.565598 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.571650 4945 scope.go:117] "RemoveContainer" containerID="ca859b837b4d6d7c6f85efa965868ddba91015b75686b509128eaf6c975919c5" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.575134 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608033 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-combined-ca-bundle\") pod \"3a31495b-7caf-4037-87b3-c0d7f29bf582\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608125 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-swift-storage-0\") pod \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608182 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w2f9\" (UniqueName: \"kubernetes.io/projected/3a31495b-7caf-4037-87b3-c0d7f29bf582-kube-api-access-5w2f9\") pod \"3a31495b-7caf-4037-87b3-c0d7f29bf582\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608264 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-nova-novncproxy-tls-certs\") pod \"3a31495b-7caf-4037-87b3-c0d7f29bf582\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608318 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-config\") pod \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608357 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-config-data\") pod \"3a31495b-7caf-4037-87b3-c0d7f29bf582\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608377 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-sb\") pod \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608406 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-nb\") pod \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608433 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-svc\") pod \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608493 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-vencrypt-tls-certs\") pod \"3a31495b-7caf-4037-87b3-c0d7f29bf582\" (UID: \"3a31495b-7caf-4037-87b3-c0d7f29bf582\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.608532 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpch8\" (UniqueName: \"kubernetes.io/projected/384f1bb0-ff94-4fc4-b658-7e1a250945ae-kube-api-access-dpch8\") pod \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\" (UID: \"384f1bb0-ff94-4fc4-b658-7e1a250945ae\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.646880 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/384f1bb0-ff94-4fc4-b658-7e1a250945ae-kube-api-access-dpch8" (OuterVolumeSpecName: "kube-api-access-dpch8") pod "384f1bb0-ff94-4fc4-b658-7e1a250945ae" (UID: "384f1bb0-ff94-4fc4-b658-7e1a250945ae"). InnerVolumeSpecName "kube-api-access-dpch8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.648336 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a31495b-7caf-4037-87b3-c0d7f29bf582-kube-api-access-5w2f9" (OuterVolumeSpecName: "kube-api-access-5w2f9") pod "3a31495b-7caf-4037-87b3-c0d7f29bf582" (UID: "3a31495b-7caf-4037-87b3-c0d7f29bf582"). InnerVolumeSpecName "kube-api-access-5w2f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.663345 4945 scope.go:117] "RemoveContainer" containerID="7e2f33a76160a89b2b209c529e605d89f63fe75180ef0858d4423cb32db2c20d" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.710017 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-etc-swift\") pod \"c64f8a71-9ee5-4145-b635-bd282dabd781\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.710249 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-combined-ca-bundle\") pod \"c64f8a71-9ee5-4145-b635-bd282dabd781\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.710353 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-public-tls-certs\") pod \"c64f8a71-9ee5-4145-b635-bd282dabd781\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.710412 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-config-data\") pod \"c64f8a71-9ee5-4145-b635-bd282dabd781\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.710433 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnr6p\" (UniqueName: \"kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-kube-api-access-jnr6p\") pod \"c64f8a71-9ee5-4145-b635-bd282dabd781\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.713133 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-log-httpd\") pod \"c64f8a71-9ee5-4145-b635-bd282dabd781\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.713174 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-internal-tls-certs\") pod \"c64f8a71-9ee5-4145-b635-bd282dabd781\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.713287 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-run-httpd\") pod \"c64f8a71-9ee5-4145-b635-bd282dabd781\" (UID: \"c64f8a71-9ee5-4145-b635-bd282dabd781\") " Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.713779 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c64f8a71-9ee5-4145-b635-bd282dabd781" (UID: "c64f8a71-9ee5-4145-b635-bd282dabd781"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.715334 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpch8\" (UniqueName: \"kubernetes.io/projected/384f1bb0-ff94-4fc4-b658-7e1a250945ae-kube-api-access-dpch8\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.715438 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.715556 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w2f9\" (UniqueName: \"kubernetes.io/projected/3a31495b-7caf-4037-87b3-c0d7f29bf582-kube-api-access-5w2f9\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.718268 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c64f8a71-9ee5-4145-b635-bd282dabd781" (UID: "c64f8a71-9ee5-4145-b635-bd282dabd781"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.728706 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapibadc-account-delete-4n95m"] Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.760967 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-kube-api-access-jnr6p" (OuterVolumeSpecName: "kube-api-access-jnr6p") pod "c64f8a71-9ee5-4145-b635-bd282dabd781" (UID: "c64f8a71-9ee5-4145-b635-bd282dabd781"). InnerVolumeSpecName "kube-api-access-jnr6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.774672 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c64f8a71-9ee5-4145-b635-bd282dabd781" (UID: "c64f8a71-9ee5-4145-b635-bd282dabd781"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.818773 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c64f8a71-9ee5-4145-b635-bd282dabd781-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.818806 4945 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.818815 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnr6p\" (UniqueName: \"kubernetes.io/projected/c64f8a71-9ee5-4145-b635-bd282dabd781-kube-api-access-jnr6p\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.973777 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "384f1bb0-ff94-4fc4-b658-7e1a250945ae" (UID: "384f1bb0-ff94-4fc4-b658-7e1a250945ae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.980700 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a31495b-7caf-4037-87b3-c0d7f29bf582" (UID: "3a31495b-7caf-4037-87b3-c0d7f29bf582"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:49 crc kubenswrapper[4945]: I1206 07:20:49.990007 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "384f1bb0-ff94-4fc4-b658-7e1a250945ae" (UID: "384f1bb0-ff94-4fc4-b658-7e1a250945ae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.026214 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-config" (OuterVolumeSpecName: "config") pod "384f1bb0-ff94-4fc4-b658-7e1a250945ae" (UID: "384f1bb0-ff94-4fc4-b658-7e1a250945ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.026336 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.026359 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.026370 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.027515 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-config-data" (OuterVolumeSpecName: "config-data") pod "3a31495b-7caf-4037-87b3-c0d7f29bf582" (UID: "3a31495b-7caf-4037-87b3-c0d7f29bf582"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.056664 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "3a31495b-7caf-4037-87b3-c0d7f29bf582" (UID: "3a31495b-7caf-4037-87b3-c0d7f29bf582"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.058059 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "384f1bb0-ff94-4fc4-b658-7e1a250945ae" (UID: "384f1bb0-ff94-4fc4-b658-7e1a250945ae"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.081167 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "384f1bb0-ff94-4fc4-b658-7e1a250945ae" (UID: "384f1bb0-ff94-4fc4-b658-7e1a250945ae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.140590 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.140627 4945 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.140641 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.140651 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.140661 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/384f1bb0-ff94-4fc4-b658-7e1a250945ae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.143397 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c64f8a71-9ee5-4145-b635-bd282dabd781" (UID: "c64f8a71-9ee5-4145-b635-bd282dabd781"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.144982 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c64f8a71-9ee5-4145-b635-bd282dabd781" (UID: "c64f8a71-9ee5-4145-b635-bd282dabd781"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.156356 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-config-data" (OuterVolumeSpecName: "config-data") pod "c64f8a71-9ee5-4145-b635-bd282dabd781" (UID: "c64f8a71-9ee5-4145-b635-bd282dabd781"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.169266 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "3a31495b-7caf-4037-87b3-c0d7f29bf582" (UID: "3a31495b-7caf-4037-87b3-c0d7f29bf582"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.175797 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c64f8a71-9ee5-4145-b635-bd282dabd781" (UID: "c64f8a71-9ee5-4145-b635-bd282dabd781"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.182493 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.168:8776/healthcheck\": read tcp 10.217.0.2:40190->10.217.0.168:8776: read: connection reset by peer" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.204562 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapibadc-account-delete-4n95m" event={"ID":"83eee8a2-f458-4193-8023-ca1228dc92a3","Type":"ContainerStarted","Data":"e1b7abd787f3f862aa8ae56e6c3ece46e8021c2d3eb8f17e25be9eec176571dd"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.216352 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0216c-account-delete-bx5zs" event={"ID":"a6fbf826-c8e6-4231-b479-90fe7de5a175","Type":"ContainerStarted","Data":"2aceb45c5bf7d83622fb1a78f07382add9a1018ebff2cef376918ef0889ca602"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.218356 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.218723 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a31495b-7caf-4037-87b3-c0d7f29bf582","Type":"ContainerDied","Data":"4d06076acad515ac5126efa1ff728044607ce8bccc093f1dec2556b021c8f8af"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.234647 4945 generic.go:334] "Generic (PLEG): container finished" podID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" containerID="14269f510b2f262408ff083c20bf00b7f68b48a2e3c497d57f6d2ea6e8cc4fcb" exitCode=0 Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.234770 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b6d74c1f-98d7-40ac-a1e8-233604971bc9","Type":"ContainerDied","Data":"14269f510b2f262408ff083c20bf00b7f68b48a2e3c497d57f6d2ea6e8cc4fcb"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.234808 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b6d74c1f-98d7-40ac-a1e8-233604971bc9","Type":"ContainerDied","Data":"499f462096d17366e60f722095d02f3e598561067542dcee4e671d295dc8df3f"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.234827 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="499f462096d17366e60f722095d02f3e598561067542dcee4e671d295dc8df3f" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.237165 4945 generic.go:334] "Generic (PLEG): container finished" podID="95dc303d-9736-4c29-a1f4-f8b362179eeb" containerID="ca5920be64a83d38ca04c8eb4934c935f589ac9dbbb41a920be90b450c3ed0c1" exitCode=0 Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.237242 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement73b1-account-delete-l54br" event={"ID":"95dc303d-9736-4c29-a1f4-f8b362179eeb","Type":"ContainerDied","Data":"ca5920be64a83d38ca04c8eb4934c935f589ac9dbbb41a920be90b450c3ed0c1"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.245598 4945 generic.go:334] "Generic (PLEG): container finished" podID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" containerID="bb3773d5b61628e5effd422bfa3be33b5be1e9ba5f96b66980463e518f4da6ac" exitCode=0 Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.245665 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ae8dae45-f01e-4ef0-8814-07511ff2fcad","Type":"ContainerDied","Data":"bb3773d5b61628e5effd422bfa3be33b5be1e9ba5f96b66980463e518f4da6ac"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.248207 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.248259 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.248333 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.248350 4945 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a31495b-7caf-4037-87b3-c0d7f29bf582-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.248361 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c64f8a71-9ee5-4145-b635-bd282dabd781-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.253672 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron975b-account-delete-nvsg5" event={"ID":"4231a0fc-b0c8-48da-8fab-75cd817dd3ef","Type":"ContainerStarted","Data":"3656e7d4d7212ac866e8c54991cfcaf41873338da41de2cded7fb37a11f86870"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.262691 4945 generic.go:334] "Generic (PLEG): container finished" podID="c1cb89b2-7918-457d-a36a-2cec380f3854" containerID="ce020a3a06ca7b55178dff84f8b51d68fe50cb6988fa5275e3943fba1bd14855" exitCode=0 Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.262759 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancef244-account-delete-vtwn6" event={"ID":"c1cb89b2-7918-457d-a36a-2cec380f3854","Type":"ContainerDied","Data":"ce020a3a06ca7b55178dff84f8b51d68fe50cb6988fa5275e3943fba1bd14855"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.280734 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican4ffa-account-delete-r98d4" event={"ID":"5e3c48ac-c191-4478-877e-9600b6f63929","Type":"ContainerStarted","Data":"182ff6178fad99f1d5b03a67029019bd53de2d67c4dfbf7f4e430acffa8b1cce"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.286508 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6ccd85df7-25dfh" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.286604 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6ccd85df7-25dfh" event={"ID":"c64f8a71-9ee5-4145-b635-bd282dabd781","Type":"ContainerDied","Data":"db08e54665d085bc316da8216960b9336d36d7829250120b04981c2d5544f1bd"} Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.307497 4945 generic.go:334] "Generic (PLEG): container finished" podID="cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b" containerID="9182d14993c72dff3cfd16a7cb6486a5a13bfa7bb047b47d0805ca2dbeb0e000" exitCode=0 Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.307608 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-n747t" Dec 06 07:20:50 crc kubenswrapper[4945]: I1206 07:20:50.309378 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder3788-account-delete-56qf8" event={"ID":"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b","Type":"ContainerDied","Data":"9182d14993c72dff3cfd16a7cb6486a5a13bfa7bb047b47d0805ca2dbeb0e000"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.673551 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.677991 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="ceilometer-central-agent" containerID="cri-o://18675bd130adeb9a967975c3bf8070d18d7c2f22354b8443abed915b5b687156" gracePeriod=30 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.678519 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="proxy-httpd" containerID="cri-o://adc583ba09127799a295f5c8a757b2e8cd330c17f41b39958b390affbc31d557" gracePeriod=30 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.678566 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="sg-core" containerID="cri-o://9e0f28c738c6aecbb802ee0805c8bd63678cb98319f526693685a2598a4babc5" gracePeriod=30 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.678606 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="ceilometer-notification-agent" containerID="cri-o://b9f8060a9b781ba6b1a59825c6f03e6c1703a2845c803ebc8a4b8e3e321741c0" gracePeriod=30 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.708403 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.708959 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d5228457-326b-4470-ad26-f7bba1f50e18" containerName="kube-state-metrics" containerID="cri-o://704c83d0821804990d78d0bdd43784abbc4a50001e79e73a668598b93f18f6ac" gracePeriod=30 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.838860 4945 scope.go:117] "RemoveContainer" containerID="89f2192e8e4d5ac29a2fb860d1dd70435b3661183bcc6f14f8a20a90fab12155" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.844968 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.845209 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" containerName="memcached" containerID="cri-o://ac89c0e1b702d43f1a10ceb5d1fa331abe2da4bec10b71837cf55f775bb4f9c7" gracePeriod=30 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.846256 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:46284->10.217.0.202:8775: read: connection reset by peer" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:50.846559 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:46272->10.217.0.202:8775: read: connection reset by peer" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.138163 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.140468 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.141113 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.141227 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.173361 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.177818 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.195574 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.195651 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovs-vswitchd" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.248660 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f9668d888-8gtj6" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.248934 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f9668d888-8gtj6" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.268636 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="003bd22c-19f6-4bfa-8f8d-3045dfb7c841" path="/var/lib/kubelet/pods/003bd22c-19f6-4bfa-8f8d-3045dfb7c841/volumes" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.269765 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b51cdd1a-131a-4f4c-8521-93890fd48722" path="/var/lib/kubelet/pods/b51cdd1a-131a-4f4c-8521-93890fd48722/volumes" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.270775 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" path="/var/lib/kubelet/pods/d5496b71-8389-4d16-97c0-5752a2f1ef2a/volumes" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272316 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2b2jk"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272352 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2b2jk"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272368 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-r6cs4"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272379 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-r6cs4"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272388 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-694b7fd487-99w9l"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272403 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystonefa31-account-delete-dt4r7"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.272693 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003bd22c-19f6-4bfa-8f8d-3045dfb7c841" containerName="openstack-network-exporter" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272706 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="003bd22c-19f6-4bfa-8f8d-3045dfb7c841" containerName="openstack-network-exporter" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.272723 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" containerName="openstack-network-exporter" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272730 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" containerName="openstack-network-exporter" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.272746 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384f1bb0-ff94-4fc4-b658-7e1a250945ae" containerName="dnsmasq-dns" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272752 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="384f1bb0-ff94-4fc4-b658-7e1a250945ae" containerName="dnsmasq-dns" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.272760 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c64f8a71-9ee5-4145-b635-bd282dabd781" containerName="proxy-httpd" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272766 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c64f8a71-9ee5-4145-b635-bd282dabd781" containerName="proxy-httpd" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.272779 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c64f8a71-9ee5-4145-b635-bd282dabd781" containerName="proxy-server" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272784 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c64f8a71-9ee5-4145-b635-bd282dabd781" containerName="proxy-server" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.272805 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a31495b-7caf-4037-87b3-c0d7f29bf582" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272812 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a31495b-7caf-4037-87b3-c0d7f29bf582" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.272821 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51cdd1a-131a-4f4c-8521-93890fd48722" containerName="openstack-network-exporter" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272829 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51cdd1a-131a-4f4c-8521-93890fd48722" containerName="openstack-network-exporter" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.272844 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384f1bb0-ff94-4fc4-b658-7e1a250945ae" containerName="init" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272850 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="384f1bb0-ff94-4fc4-b658-7e1a250945ae" containerName="init" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.272861 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51cdd1a-131a-4f4c-8521-93890fd48722" containerName="ovsdbserver-sb" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272867 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51cdd1a-131a-4f4c-8521-93890fd48722" containerName="ovsdbserver-sb" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.272878 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" containerName="ovsdbserver-nb" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.272884 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" containerName="ovsdbserver-nb" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.273039 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="384f1bb0-ff94-4fc4-b658-7e1a250945ae" containerName="dnsmasq-dns" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.273053 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51cdd1a-131a-4f4c-8521-93890fd48722" containerName="openstack-network-exporter" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.273063 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="003bd22c-19f6-4bfa-8f8d-3045dfb7c841" containerName="openstack-network-exporter" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.273070 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" containerName="ovsdbserver-nb" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.273077 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5496b71-8389-4d16-97c0-5752a2f1ef2a" containerName="openstack-network-exporter" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.273084 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a31495b-7caf-4037-87b3-c0d7f29bf582" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.273097 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c64f8a71-9ee5-4145-b635-bd282dabd781" containerName="proxy-httpd" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.273106 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51cdd1a-131a-4f4c-8521-93890fd48722" containerName="ovsdbserver-sb" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.273115 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c64f8a71-9ee5-4145-b635-bd282dabd781" containerName="proxy-server" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.283176 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-694b7fd487-99w9l" podUID="5af85716-fd7e-45e9-8320-4db8800302a8" containerName="keystone-api" containerID="cri-o://9d35634fe6ffb6779224176c397d8848048afbbeb060d73a9ed0f8871a1cd24b" gracePeriod=30 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295121 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystonefa31-account-delete-dt4r7"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295157 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295174 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-zxg5j"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295192 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-zxg5j"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295216 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-216c-account-create-update-gvcn7"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295234 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0216c-account-delete-bx5zs"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295250 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-216c-account-create-update-gvcn7"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295260 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-m5ptv"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295271 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-m5ptv"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295297 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystonefa31-account-delete-dt4r7"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295307 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-fa31-account-create-update-dbjwp"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295319 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-fa31-account-create-update-dbjwp"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295328 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-lstm6"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295338 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-lstm6"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295349 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-badc-account-create-update-njdjh"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295359 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapibadc-account-delete-4n95m"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.295725 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonefa31-account-delete-dt4r7" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.303574 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-badc-account-create-update-njdjh"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.308349 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5984506921a2b100cf94ec4a7e72d52586d9d11cdd69b262d6ee89622fc5fcf4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.322575 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5984506921a2b100cf94ec4a7e72d52586d9d11cdd69b262d6ee89622fc5fcf4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.334925 4945 generic.go:334] "Generic (PLEG): container finished" podID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerID="b957bb81cfc39bf3a8747b1e1804ba1ec5dcb33d0889082d694cf4ffd810ef17" exitCode=0 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.334996 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"edbffd82-c356-43a3-94a3-3549b7d9cd31","Type":"ContainerDied","Data":"b957bb81cfc39bf3a8747b1e1804ba1ec5dcb33d0889082d694cf4ffd810ef17"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.335033 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"edbffd82-c356-43a3-94a3-3549b7d9cd31","Type":"ContainerDied","Data":"7649c561156cfdd3269bbf7b7920020626421572c2eafcd315a0c2aebc66ac69"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.335045 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7649c561156cfdd3269bbf7b7920020626421572c2eafcd315a0c2aebc66ac69" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.336102 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5984506921a2b100cf94ec4a7e72d52586d9d11cdd69b262d6ee89622fc5fcf4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.336137 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6ae75a6a-2656-4e65-a84d-8767a570ce8f" containerName="nova-scheduler-scheduler" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.338717 4945 generic.go:334] "Generic (PLEG): container finished" podID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerID="a7da8e3a50d0a0f1e3ae039655940701da64854727b798ffef58b2a42ba05b35" exitCode=0 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.338772 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"efad5b26-d064-4d06-8cc8-361d094dfa9c","Type":"ContainerDied","Data":"a7da8e3a50d0a0f1e3ae039655940701da64854727b798ffef58b2a42ba05b35"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.354728 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xc2dd" podUID="3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" containerName="ovn-controller" probeResult="failure" output=< Dec 06 07:20:51 crc kubenswrapper[4945]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Dec 06 07:20:51 crc kubenswrapper[4945]: > Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.368665 4945 generic.go:334] "Generic (PLEG): container finished" podID="d5228457-326b-4470-ad26-f7bba1f50e18" containerID="704c83d0821804990d78d0bdd43784abbc4a50001e79e73a668598b93f18f6ac" exitCode=2 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.368787 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d5228457-326b-4470-ad26-f7bba1f50e18","Type":"ContainerDied","Data":"704c83d0821804990d78d0bdd43784abbc4a50001e79e73a668598b93f18f6ac"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.387490 4945 generic.go:334] "Generic (PLEG): container finished" podID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" containerID="f75d79cb35ec6aeed2fbb5df193689624fa68ec6f36d6f07600634a6a1eb3d4f" exitCode=0 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.387611 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ba3e057-d321-48ca-b7a0-8d791cbb7169","Type":"ContainerDied","Data":"f75d79cb35ec6aeed2fbb5df193689624fa68ec6f36d6f07600634a6a1eb3d4f"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.405517 4945 generic.go:334] "Generic (PLEG): container finished" podID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerID="adc583ba09127799a295f5c8a757b2e8cd330c17f41b39958b390affbc31d557" exitCode=0 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.405561 4945 generic.go:334] "Generic (PLEG): container finished" podID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerID="9e0f28c738c6aecbb802ee0805c8bd63678cb98319f526693685a2598a4babc5" exitCode=2 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.405656 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8fbb8a7-4226-4b2f-9e95-621059e6283d","Type":"ContainerDied","Data":"adc583ba09127799a295f5c8a757b2e8cd330c17f41b39958b390affbc31d557"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.405692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8fbb8a7-4226-4b2f-9e95-621059e6283d","Type":"ContainerDied","Data":"9e0f28c738c6aecbb802ee0805c8bd63678cb98319f526693685a2598a4babc5"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.414955 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.417510 4945 generic.go:334] "Generic (PLEG): container finished" podID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" containerID="f1978a8cc52c8c7435676004e144a0b8373ccf11fd0f8cada610a31d96b4bf20" exitCode=0 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.417575 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf0fa4a-4a9c-4644-9fe3-413500c59a98","Type":"ContainerDied","Data":"f1978a8cc52c8c7435676004e144a0b8373ccf11fd0f8cada610a31d96b4bf20"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.430876 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerID="afba2827fdc58ee118d91cef305ad6d598b1747d9ddf4810bfde53456fdf7417" exitCode=0 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.430963 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9668d888-8gtj6" event={"ID":"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975","Type":"ContainerDied","Data":"afba2827fdc58ee118d91cef305ad6d598b1747d9ddf4810bfde53456fdf7417"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.445068 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e3c48ac-c191-4478-877e-9600b6f63929" containerID="182ff6178fad99f1d5b03a67029019bd53de2d67c4dfbf7f4e430acffa8b1cce" exitCode=0 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.445178 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican4ffa-account-delete-r98d4" event={"ID":"5e3c48ac-c191-4478-877e-9600b6f63929","Type":"ContainerDied","Data":"182ff6178fad99f1d5b03a67029019bd53de2d67c4dfbf7f4e430acffa8b1cce"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.456553 4945 generic.go:334] "Generic (PLEG): container finished" podID="569f0e1c-e58f-4e87-bde3-e7be4659639a" containerID="54e516865e0c2bf291e50817a1f0a0df2b67a1322700ab4cf577d2385c08bcb2" exitCode=0 Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.456613 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75599744c6-dmqzn" event={"ID":"569f0e1c-e58f-4e87-bde3-e7be4659639a","Type":"ContainerDied","Data":"54e516865e0c2bf291e50817a1f0a0df2b67a1322700ab4cf577d2385c08bcb2"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.456650 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75599744c6-dmqzn" event={"ID":"569f0e1c-e58f-4e87-bde3-e7be4659639a","Type":"ContainerDied","Data":"a4d123d133c5818a158baecbca329d7d1fd35c32d80f99170381689fd8534f7e"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.456663 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4d123d133c5818a158baecbca329d7d1fd35c32d80f99170381689fd8534f7e" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.458172 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ae8dae45-f01e-4ef0-8814-07511ff2fcad","Type":"ContainerDied","Data":"4a4f53aad4d8b8ad07a84460aeafda04a68f1d5b681c2c2d550714a3c60a4766"} Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.458188 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a4f53aad4d8b8ad07a84460aeafda04a68f1d5b681c2c2d550714a3c60a4766" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.549795 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51e0e29bf34bea465946d1ff2713e5af788a26b0c2039d1356719fdaad5afb73" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.559600 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51e0e29bf34bea465946d1ff2713e5af788a26b0c2039d1356719fdaad5afb73" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.561684 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51e0e29bf34bea465946d1ff2713e5af788a26b0c2039d1356719fdaad5afb73" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.561738 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="22eb2234-b14b-474c-b1bf-cf726acb064b" containerName="nova-cell1-conductor-conductor" Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.718518 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fd592672f76730e847191d2934d9c9cfb67d21204f6a000e593a960e8df07a85" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.720646 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fd592672f76730e847191d2934d9c9cfb67d21204f6a000e593a960e8df07a85" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.721868 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fd592672f76730e847191d2934d9c9cfb67d21204f6a000e593a960e8df07a85" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 07:20:51 crc kubenswrapper[4945]: E1206 07:20:51.721900 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="72607de1-4f23-441e-b722-3844b2eb6d15" containerName="nova-cell0-conductor-conductor" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.727922 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="250c85e2-e14d-4f93-822c-c9a88352eafe" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Dec 06 07:20:51 crc kubenswrapper[4945]: I1206 07:20:51.767601 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="fe30cfa7-9746-4771-bac5-29e3a3662aec" containerName="galera" containerID="cri-o://bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8" gracePeriod=30 Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.293539 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.315932 4945 scope.go:117] "RemoveContainer" containerID="29a489e1a3fc4b892c014063d26ad8a5a7da41074019dcb533e94027b8e256f2" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.355013 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data-custom\") pod \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.355049 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-combined-ca-bundle\") pod \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.355163 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-scripts\") pod \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.355209 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data\") pod \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.355241 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkmkq\" (UniqueName: \"kubernetes.io/projected/b6d74c1f-98d7-40ac-a1e8-233604971bc9-kube-api-access-mkmkq\") pod \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.355320 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d74c1f-98d7-40ac-a1e8-233604971bc9-etc-machine-id\") pod \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\" (UID: \"b6d74c1f-98d7-40ac-a1e8-233604971bc9\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.355721 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6d74c1f-98d7-40ac-a1e8-233604971bc9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b6d74c1f-98d7-40ac-a1e8-233604971bc9" (UID: "b6d74c1f-98d7-40ac-a1e8-233604971bc9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.370467 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-scripts" (OuterVolumeSpecName: "scripts") pod "b6d74c1f-98d7-40ac-a1e8-233604971bc9" (UID: "b6d74c1f-98d7-40ac-a1e8-233604971bc9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.391469 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.414139 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b6d74c1f-98d7-40ac-a1e8-233604971bc9" (UID: "b6d74c1f-98d7-40ac-a1e8-233604971bc9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.419377 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.420733 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d74c1f-98d7-40ac-a1e8-233604971bc9-kube-api-access-mkmkq" (OuterVolumeSpecName: "kube-api-access-mkmkq") pod "b6d74c1f-98d7-40ac-a1e8-233604971bc9" (UID: "b6d74c1f-98d7-40ac-a1e8-233604971bc9"). InnerVolumeSpecName "kube-api-access-mkmkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.420878 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonefa31-account-delete-dt4r7" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.421067 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.425547 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.438080 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.444872 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457022 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-default\") pod \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457087 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457129 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-combined-ca-bundle\") pod \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457162 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-internal-tls-certs\") pod \"569f0e1c-e58f-4e87-bde3-e7be4659639a\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457199 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-operator-scripts\") pod \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457233 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data-custom\") pod \"edbffd82-c356-43a3-94a3-3549b7d9cd31\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457267 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfx87\" (UniqueName: \"kubernetes.io/projected/569f0e1c-e58f-4e87-bde3-e7be4659639a-kube-api-access-lfx87\") pod \"569f0e1c-e58f-4e87-bde3-e7be4659639a\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457886 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-scripts\") pod \"569f0e1c-e58f-4e87-bde3-e7be4659639a\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457914 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-combined-ca-bundle\") pod \"569f0e1c-e58f-4e87-bde3-e7be4659639a\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457934 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-combined-ca-bundle\") pod \"edbffd82-c356-43a3-94a3-3549b7d9cd31\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457963 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-logs\") pod \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.457983 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kolla-config\") pod \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458033 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-config-data\") pod \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458060 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-internal-tls-certs\") pod \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458079 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458100 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edbffd82-c356-43a3-94a3-3549b7d9cd31-etc-machine-id\") pod \"edbffd82-c356-43a3-94a3-3549b7d9cd31\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458133 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-generated\") pod \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458149 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edbffd82-c356-43a3-94a3-3549b7d9cd31-logs\") pod \"edbffd82-c356-43a3-94a3-3549b7d9cd31\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458168 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-scripts\") pod \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458206 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44pbb\" (UniqueName: \"kubernetes.io/projected/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kube-api-access-44pbb\") pod \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458222 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-scripts\") pod \"edbffd82-c356-43a3-94a3-3549b7d9cd31\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458241 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-httpd-run\") pod \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458260 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/569f0e1c-e58f-4e87-bde3-e7be4659639a-logs\") pod \"569f0e1c-e58f-4e87-bde3-e7be4659639a\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458300 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-public-tls-certs\") pod \"569f0e1c-e58f-4e87-bde3-e7be4659639a\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458321 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-galera-tls-certs\") pod \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458336 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-combined-ca-bundle\") pod \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\" (UID: \"ae8dae45-f01e-4ef0-8814-07511ff2fcad\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458368 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m595p\" (UniqueName: \"kubernetes.io/projected/edbffd82-c356-43a3-94a3-3549b7d9cd31-kube-api-access-m595p\") pod \"edbffd82-c356-43a3-94a3-3549b7d9cd31\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.458412 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-internal-tls-certs\") pod \"edbffd82-c356-43a3-94a3-3549b7d9cd31\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.459268 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0bf0fa4a-4a9c-4644-9fe3-413500c59a98" (UID: "0bf0fa4a-4a9c-4644-9fe3-413500c59a98"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.459327 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/edbffd82-c356-43a3-94a3-3549b7d9cd31-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "edbffd82-c356-43a3-94a3-3549b7d9cd31" (UID: "edbffd82-c356-43a3-94a3-3549b7d9cd31"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.466452 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "ae8dae45-f01e-4ef0-8814-07511ff2fcad" (UID: "ae8dae45-f01e-4ef0-8814-07511ff2fcad"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.469541 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edbffd82-c356-43a3-94a3-3549b7d9cd31-logs" (OuterVolumeSpecName: "logs") pod "edbffd82-c356-43a3-94a3-3549b7d9cd31" (UID: "edbffd82-c356-43a3-94a3-3549b7d9cd31"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.471938 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6ccd85df7-25dfh"] Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.474420 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "ae8dae45-f01e-4ef0-8814-07511ff2fcad" (UID: "ae8dae45-f01e-4ef0-8814-07511ff2fcad"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.475553 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/569f0e1c-e58f-4e87-bde3-e7be4659639a-logs" (OuterVolumeSpecName: "logs") pod "569f0e1c-e58f-4e87-bde3-e7be4659639a" (UID: "569f0e1c-e58f-4e87-bde3-e7be4659639a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.471406 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljgs5\" (UniqueName: \"kubernetes.io/projected/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-kube-api-access-ljgs5\") pod \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\" (UID: \"0bf0fa4a-4a9c-4644-9fe3-413500c59a98\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.480608 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data\") pod \"edbffd82-c356-43a3-94a3-3549b7d9cd31\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.480666 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-public-tls-certs\") pod \"edbffd82-c356-43a3-94a3-3549b7d9cd31\" (UID: \"edbffd82-c356-43a3-94a3-3549b7d9cd31\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.480697 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-config-data\") pod \"569f0e1c-e58f-4e87-bde3-e7be4659639a\" (UID: \"569f0e1c-e58f-4e87-bde3-e7be4659639a\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.481418 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.481431 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edbffd82-c356-43a3-94a3-3549b7d9cd31-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.481440 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edbffd82-c356-43a3-94a3-3549b7d9cd31-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.481453 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.481463 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.481471 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.481478 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/569f0e1c-e58f-4e87-bde3-e7be4659639a-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.481486 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkmkq\" (UniqueName: \"kubernetes.io/projected/b6d74c1f-98d7-40ac-a1e8-233604971bc9-kube-api-access-mkmkq\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.481495 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d74c1f-98d7-40ac-a1e8-233604971bc9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.481503 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: E1206 07:20:52.481569 4945 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 06 07:20:52 crc kubenswrapper[4945]: E1206 07:20:52.481629 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data podName:250c85e2-e14d-4f93-822c-c9a88352eafe nodeName:}" failed. No retries permitted until 2025-12-06 07:21:00.481607272 +0000 UTC m=+1673.936468316 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data") pod "rabbitmq-cell1-server-0" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe") : configmap "rabbitmq-cell1-config-data" not found Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.488874 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae8dae45-f01e-4ef0-8814-07511ff2fcad" (UID: "ae8dae45-f01e-4ef0-8814-07511ff2fcad"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.489691 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "ae8dae45-f01e-4ef0-8814-07511ff2fcad" (UID: "ae8dae45-f01e-4ef0-8814-07511ff2fcad"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.490084 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-logs" (OuterVolumeSpecName: "logs") pod "0bf0fa4a-4a9c-4644-9fe3-413500c59a98" (UID: "0bf0fa4a-4a9c-4644-9fe3-413500c59a98"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.500374 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edbffd82-c356-43a3-94a3-3549b7d9cd31-kube-api-access-m595p" (OuterVolumeSpecName: "kube-api-access-m595p") pod "edbffd82-c356-43a3-94a3-3549b7d9cd31" (UID: "edbffd82-c356-43a3-94a3-3549b7d9cd31"). InnerVolumeSpecName "kube-api-access-m595p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.501633 4945 scope.go:117] "RemoveContainer" containerID="6afab1fa21766e8cbfdedb4c1868e051427546f3e9b764dfcadfaac9ac9f3aa3" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.548555 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-6ccd85df7-25dfh"] Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.554907 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.563911 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-n747t"] Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.571556 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kube-api-access-44pbb" (OuterVolumeSpecName: "kube-api-access-44pbb") pod "ae8dae45-f01e-4ef0-8814-07511ff2fcad" (UID: "ae8dae45-f01e-4ef0-8814-07511ff2fcad"). InnerVolumeSpecName "kube-api-access-44pbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.573884 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-scripts" (OuterVolumeSpecName: "scripts") pod "edbffd82-c356-43a3-94a3-3549b7d9cd31" (UID: "edbffd82-c356-43a3-94a3-3549b7d9cd31"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.575331 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-n747t"] Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.589396 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.589422 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.589433 4945 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.589445 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44pbb\" (UniqueName: \"kubernetes.io/projected/ae8dae45-f01e-4ef0-8814-07511ff2fcad-kube-api-access-44pbb\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.589456 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.589469 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m595p\" (UniqueName: \"kubernetes.io/projected/edbffd82-c356-43a3-94a3-3549b7d9cd31-kube-api-access-m595p\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.615464 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-kube-api-access-ljgs5" (OuterVolumeSpecName: "kube-api-access-ljgs5") pod "0bf0fa4a-4a9c-4644-9fe3-413500c59a98" (UID: "0bf0fa4a-4a9c-4644-9fe3-413500c59a98"). InnerVolumeSpecName "kube-api-access-ljgs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.618380 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-scripts" (OuterVolumeSpecName: "scripts") pod "0bf0fa4a-4a9c-4644-9fe3-413500c59a98" (UID: "0bf0fa4a-4a9c-4644-9fe3-413500c59a98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.618562 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "0bf0fa4a-4a9c-4644-9fe3-413500c59a98" (UID: "0bf0fa4a-4a9c-4644-9fe3-413500c59a98"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.619323 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/569f0e1c-e58f-4e87-bde3-e7be4659639a-kube-api-access-lfx87" (OuterVolumeSpecName: "kube-api-access-lfx87") pod "569f0e1c-e58f-4e87-bde3-e7be4659639a" (UID: "569f0e1c-e58f-4e87-bde3-e7be4659639a"). InnerVolumeSpecName "kube-api-access-lfx87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.625529 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-scripts" (OuterVolumeSpecName: "scripts") pod "569f0e1c-e58f-4e87-bde3-e7be4659639a" (UID: "569f0e1c-e58f-4e87-bde3-e7be4659639a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.639352 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "edbffd82-c356-43a3-94a3-3549b7d9cd31" (UID: "edbffd82-c356-43a3-94a3-3549b7d9cd31"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: E1206 07:20:52.665700 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.671161 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder3788-account-delete-56qf8" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.671500 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement73b1-account-delete-l54br" Dec 06 07:20:52 crc kubenswrapper[4945]: E1206 07:20:52.672843 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.673253 4945 generic.go:334] "Generic (PLEG): container finished" podID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerID="18675bd130adeb9a967975c3bf8070d18d7c2f22354b8443abed915b5b687156" exitCode=0 Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.673346 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8fbb8a7-4226-4b2f-9e95-621059e6283d","Type":"ContainerDied","Data":"18675bd130adeb9a967975c3bf8070d18d7c2f22354b8443abed915b5b687156"} Dec 06 07:20:52 crc kubenswrapper[4945]: E1206 07:20:52.676041 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:20:52 crc kubenswrapper[4945]: E1206 07:20:52.676146 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="fe30cfa7-9746-4771-bac5-29e3a3662aec" containerName="galera" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.678034 4945 generic.go:334] "Generic (PLEG): container finished" podID="72607de1-4f23-441e-b722-3844b2eb6d15" containerID="fd592672f76730e847191d2934d9c9cfb67d21204f6a000e593a960e8df07a85" exitCode=0 Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.678107 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"72607de1-4f23-441e-b722-3844b2eb6d15","Type":"ContainerDied","Data":"fd592672f76730e847191d2934d9c9cfb67d21204f6a000e593a960e8df07a85"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.679554 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.680754 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9668d888-8gtj6" event={"ID":"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975","Type":"ContainerDied","Data":"67db1ce53a734cdc8d1cb41de803b70c6fd26d779a4527d1909b5ec2a833c239"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.683305 4945 generic.go:334] "Generic (PLEG): container finished" podID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" containerID="8f06adda9c4f122a20f297ca0d7ccd2c0349a987c76bc355daf2b4ea17945592" exitCode=0 Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.683354 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-687fd77bc7-9rms8" event={"ID":"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac","Type":"ContainerDied","Data":"8f06adda9c4f122a20f297ca0d7ccd2c0349a987c76bc355daf2b4ea17945592"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.685408 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"efad5b26-d064-4d06-8cc8-361d094dfa9c","Type":"ContainerDied","Data":"d57aa4467dcf950661689bb2640542998c0dd1f9ef3e33c8f455cbc0e88d8c6b"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.685493 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.692382 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt4bx\" (UniqueName: \"kubernetes.io/projected/efad5b26-d064-4d06-8cc8-361d094dfa9c-kube-api-access-rt4bx\") pod \"efad5b26-d064-4d06-8cc8-361d094dfa9c\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.692666 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-combined-ca-bundle\") pod \"efad5b26-d064-4d06-8cc8-361d094dfa9c\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.692723 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efad5b26-d064-4d06-8cc8-361d094dfa9c-logs\") pod \"efad5b26-d064-4d06-8cc8-361d094dfa9c\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.692750 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-config-data\") pod \"efad5b26-d064-4d06-8cc8-361d094dfa9c\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.692769 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-nova-metadata-tls-certs\") pod \"efad5b26-d064-4d06-8cc8-361d094dfa9c\" (UID: \"efad5b26-d064-4d06-8cc8-361d094dfa9c\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.694469 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljgs5\" (UniqueName: \"kubernetes.io/projected/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-kube-api-access-ljgs5\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.694488 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.694497 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfx87\" (UniqueName: \"kubernetes.io/projected/569f0e1c-e58f-4e87-bde3-e7be4659639a-kube-api-access-lfx87\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.694506 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.694524 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.699505 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.698529 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efad5b26-d064-4d06-8cc8-361d094dfa9c-logs" (OuterVolumeSpecName: "logs") pod "efad5b26-d064-4d06-8cc8-361d094dfa9c" (UID: "efad5b26-d064-4d06-8cc8-361d094dfa9c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.716648 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.716651 4945 generic.go:334] "Generic (PLEG): container finished" podID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerID="145d4313977864fb199b0a745f9e48fe66f8d8ec1a67bdd8264719366ba4588e" exitCode=0 Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.716719 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9296c3b2-7997-45a0-a7e9-bc1f45ec204e","Type":"ContainerDied","Data":"145d4313977864fb199b0a745f9e48fe66f8d8ec1a67bdd8264719366ba4588e"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.717962 4945 scope.go:117] "RemoveContainer" containerID="afba2827fdc58ee118d91cef305ad6d598b1747d9ddf4810bfde53456fdf7417" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.723625 4945 generic.go:334] "Generic (PLEG): container finished" podID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" containerID="c997070637d077fb0bd2b5c6e603ae66c9789fb87271122eac47abaf2b27318e" exitCode=0 Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.723746 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7db96d494-qst22" event={"ID":"966c9b21-65c4-4d2a-af0f-0041a9bfea08","Type":"ContainerDied","Data":"c997070637d077fb0bd2b5c6e603ae66c9789fb87271122eac47abaf2b27318e"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.725207 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancef244-account-delete-vtwn6" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.729763 4945 generic.go:334] "Generic (PLEG): container finished" podID="22eb2234-b14b-474c-b1bf-cf726acb064b" containerID="51e0e29bf34bea465946d1ff2713e5af788a26b0c2039d1356719fdaad5afb73" exitCode=0 Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.729817 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"22eb2234-b14b-474c-b1bf-cf726acb064b","Type":"ContainerDied","Data":"51e0e29bf34bea465946d1ff2713e5af788a26b0c2039d1356719fdaad5afb73"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.737893 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "mysql-db") pod "ae8dae45-f01e-4ef0-8814-07511ff2fcad" (UID: "ae8dae45-f01e-4ef0-8814-07511ff2fcad"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.738134 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican4ffa-account-delete-r98d4" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.742415 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efad5b26-d064-4d06-8cc8-361d094dfa9c-kube-api-access-rt4bx" (OuterVolumeSpecName: "kube-api-access-rt4bx") pod "efad5b26-d064-4d06-8cc8-361d094dfa9c" (UID: "efad5b26-d064-4d06-8cc8-361d094dfa9c"). InnerVolumeSpecName "kube-api-access-rt4bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.743301 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancef244-account-delete-vtwn6" event={"ID":"c1cb89b2-7918-457d-a36a-2cec380f3854","Type":"ContainerDied","Data":"6a3b2affe563ac0f25fee01855b43264336522f34295ffc38cae32aa2dd70751"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.743520 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a3b2affe563ac0f25fee01855b43264336522f34295ffc38cae32aa2dd70751" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.743342 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancef244-account-delete-vtwn6" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.764384 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bf0fa4a-4a9c-4644-9fe3-413500c59a98","Type":"ContainerDied","Data":"1f62cdaba764c253ae0c59729461040b63726374f0f788c093cebdb6d800f44a"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.764447 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.768710 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement73b1-account-delete-l54br" event={"ID":"95dc303d-9736-4c29-a1f4-f8b362179eeb","Type":"ContainerDied","Data":"9f3461d0026224bb1e1e73854fad6043ef5f472047ddc7b37ea6f67d44b0fc12"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.768743 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f3461d0026224bb1e1e73854fad6043ef5f472047ddc7b37ea6f67d44b0fc12" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.768764 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement73b1-account-delete-l54br" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.770351 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder3788-account-delete-56qf8" event={"ID":"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b","Type":"ContainerDied","Data":"db5de8009af333da3fb0800bc459c1663d4ea20f27eee067c31e33b99be107fc"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.770396 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db5de8009af333da3fb0800bc459c1663d4ea20f27eee067c31e33b99be107fc" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.770458 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder3788-account-delete-56qf8" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809383 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x55dk\" (UniqueName: \"kubernetes.io/projected/c1cb89b2-7918-457d-a36a-2cec380f3854-kube-api-access-x55dk\") pod \"c1cb89b2-7918-457d-a36a-2cec380f3854\" (UID: \"c1cb89b2-7918-457d-a36a-2cec380f3854\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809458 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95dc303d-9736-4c29-a1f4-f8b362179eeb-operator-scripts\") pod \"95dc303d-9736-4c29-a1f4-f8b362179eeb\" (UID: \"95dc303d-9736-4c29-a1f4-f8b362179eeb\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809491 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-public-tls-certs\") pod \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809618 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-certs\") pod \"d5228457-326b-4470-ad26-f7bba1f50e18\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809663 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-internal-tls-certs\") pod \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809697 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvf82\" (UniqueName: \"kubernetes.io/projected/95dc303d-9736-4c29-a1f4-f8b362179eeb-kube-api-access-tvf82\") pod \"95dc303d-9736-4c29-a1f4-f8b362179eeb\" (UID: \"95dc303d-9736-4c29-a1f4-f8b362179eeb\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809788 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-combined-ca-bundle\") pod \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809851 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-logs\") pod \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809873 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e3c48ac-c191-4478-877e-9600b6f63929-operator-scripts\") pod \"5e3c48ac-c191-4478-877e-9600b6f63929\" (UID: \"5e3c48ac-c191-4478-877e-9600b6f63929\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809916 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-combined-ca-bundle\") pod \"d5228457-326b-4470-ad26-f7bba1f50e18\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.809957 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wfp7\" (UniqueName: \"kubernetes.io/projected/5e3c48ac-c191-4478-877e-9600b6f63929-kube-api-access-5wfp7\") pod \"5e3c48ac-c191-4478-877e-9600b6f63929\" (UID: \"5e3c48ac-c191-4478-877e-9600b6f63929\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.810033 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-config\") pod \"d5228457-326b-4470-ad26-f7bba1f50e18\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.810066 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpzd7\" (UniqueName: \"kubernetes.io/projected/d5228457-326b-4470-ad26-f7bba1f50e18-kube-api-access-zpzd7\") pod \"d5228457-326b-4470-ad26-f7bba1f50e18\" (UID: \"d5228457-326b-4470-ad26-f7bba1f50e18\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.810146 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-operator-scripts\") pod \"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b\" (UID: \"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.810193 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8hw2\" (UniqueName: \"kubernetes.io/projected/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-kube-api-access-x8hw2\") pod \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.810249 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj76b\" (UniqueName: \"kubernetes.io/projected/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-kube-api-access-pj76b\") pod \"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b\" (UID: \"cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.812198 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data-custom\") pod \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.812316 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data\") pod \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\" (UID: \"5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.812378 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1cb89b2-7918-457d-a36a-2cec380f3854-operator-scripts\") pod \"c1cb89b2-7918-457d-a36a-2cec380f3854\" (UID: \"c1cb89b2-7918-457d-a36a-2cec380f3854\") " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.813712 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efad5b26-d064-4d06-8cc8-361d094dfa9c-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.813737 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt4bx\" (UniqueName: \"kubernetes.io/projected/efad5b26-d064-4d06-8cc8-361d094dfa9c-kube-api-access-rt4bx\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.813771 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.819067 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95dc303d-9736-4c29-a1f4-f8b362179eeb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95dc303d-9736-4c29-a1f4-f8b362179eeb" (UID: "95dc303d-9736-4c29-a1f4-f8b362179eeb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.830581 4945 generic.go:334] "Generic (PLEG): container finished" podID="ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" containerID="ac89c0e1b702d43f1a10ceb5d1fa331abe2da4bec10b71837cf55f775bb4f9c7" exitCode=0 Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.832863 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican4ffa-account-delete-r98d4" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.833661 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b" (UID: "cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.839221 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e3c48ac-c191-4478-877e-9600b6f63929-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5e3c48ac-c191-4478-877e-9600b6f63929" (UID: "5e3c48ac-c191-4478-877e-9600b6f63929"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.839426 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-logs" (OuterVolumeSpecName: "logs") pod "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" (UID: "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.839480 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67","Type":"ContainerDied","Data":"ac89c0e1b702d43f1a10ceb5d1fa331abe2da4bec10b71837cf55f775bb4f9c7"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.842026 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75599744c6-dmqzn" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.843164 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonefa31-account-delete-dt4r7" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.843224 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.844718 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.845525 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.845935 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.846202 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican4ffa-account-delete-r98d4" event={"ID":"5e3c48ac-c191-4478-877e-9600b6f63929","Type":"ContainerDied","Data":"16290c3dc010794f4acd169c72aa06539ee1641e2027d05f2b3c23d813040254"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.846259 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16290c3dc010794f4acd169c72aa06539ee1641e2027d05f2b3c23d813040254" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.846300 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d5228457-326b-4470-ad26-f7bba1f50e18","Type":"ContainerDied","Data":"4a66f7add2b33b5fbe9aeb2b42da28221ec0eda314e30a56ff1222f6c5b78ab3"} Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.860712 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1cb89b2-7918-457d-a36a-2cec380f3854-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c1cb89b2-7918-457d-a36a-2cec380f3854" (UID: "c1cb89b2-7918-457d-a36a-2cec380f3854"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.886423 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1cb89b2-7918-457d-a36a-2cec380f3854-kube-api-access-x55dk" (OuterVolumeSpecName: "kube-api-access-x55dk") pod "c1cb89b2-7918-457d-a36a-2cec380f3854" (UID: "c1cb89b2-7918-457d-a36a-2cec380f3854"). InnerVolumeSpecName "kube-api-access-x55dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.887097 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-kube-api-access-pj76b" (OuterVolumeSpecName: "kube-api-access-pj76b") pod "cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b" (UID: "cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b"). InnerVolumeSpecName "kube-api-access-pj76b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.887266 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-kube-api-access-x8hw2" (OuterVolumeSpecName: "kube-api-access-x8hw2") pod "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" (UID: "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975"). InnerVolumeSpecName "kube-api-access-x8hw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.888553 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5228457-326b-4470-ad26-f7bba1f50e18-kube-api-access-zpzd7" (OuterVolumeSpecName: "kube-api-access-zpzd7") pod "d5228457-326b-4470-ad26-f7bba1f50e18" (UID: "d5228457-326b-4470-ad26-f7bba1f50e18"). InnerVolumeSpecName "kube-api-access-zpzd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.888682 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" (UID: "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.914340 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95dc303d-9736-4c29-a1f4-f8b362179eeb-kube-api-access-tvf82" (OuterVolumeSpecName: "kube-api-access-tvf82") pod "95dc303d-9736-4c29-a1f4-f8b362179eeb" (UID: "95dc303d-9736-4c29-a1f4-f8b362179eeb"). InnerVolumeSpecName "kube-api-access-tvf82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.914429 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e3c48ac-c191-4478-877e-9600b6f63929-kube-api-access-5wfp7" (OuterVolumeSpecName: "kube-api-access-5wfp7") pod "5e3c48ac-c191-4478-877e-9600b6f63929" (UID: "5e3c48ac-c191-4478-877e-9600b6f63929"). InnerVolumeSpecName "kube-api-access-5wfp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922240 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922270 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8hw2\" (UniqueName: \"kubernetes.io/projected/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-kube-api-access-x8hw2\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922295 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj76b\" (UniqueName: \"kubernetes.io/projected/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b-kube-api-access-pj76b\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922306 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922315 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1cb89b2-7918-457d-a36a-2cec380f3854-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922326 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x55dk\" (UniqueName: \"kubernetes.io/projected/c1cb89b2-7918-457d-a36a-2cec380f3854-kube-api-access-x55dk\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922335 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95dc303d-9736-4c29-a1f4-f8b362179eeb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922344 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvf82\" (UniqueName: \"kubernetes.io/projected/95dc303d-9736-4c29-a1f4-f8b362179eeb-kube-api-access-tvf82\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922354 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922364 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e3c48ac-c191-4478-877e-9600b6f63929-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922376 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wfp7\" (UniqueName: \"kubernetes.io/projected/5e3c48ac-c191-4478-877e-9600b6f63929-kube-api-access-5wfp7\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.922385 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpzd7\" (UniqueName: \"kubernetes.io/projected/d5228457-326b-4470-ad26-f7bba1f50e18-kube-api-access-zpzd7\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.974722 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f52cc49-cbae-487e-8d29-bdcb20fa6ddd" path="/var/lib/kubelet/pods/0f52cc49-cbae-487e-8d29-bdcb20fa6ddd/volumes" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.976234 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18f99e6e-1055-4fef-b134-e49c03a0e06a" path="/var/lib/kubelet/pods/18f99e6e-1055-4fef-b134-e49c03a0e06a/volumes" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.977118 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="384f1bb0-ff94-4fc4-b658-7e1a250945ae" path="/var/lib/kubelet/pods/384f1bb0-ff94-4fc4-b658-7e1a250945ae/volumes" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.978160 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a31495b-7caf-4037-87b3-c0d7f29bf582" path="/var/lib/kubelet/pods/3a31495b-7caf-4037-87b3-c0d7f29bf582/volumes" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.978714 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2897e5-a177-4cb5-9790-4fa9bbb35104" path="/var/lib/kubelet/pods/5e2897e5-a177-4cb5-9790-4fa9bbb35104/volumes" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.979217 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce3702d-5de0-469b-a8e8-b531c4ea5afa" path="/var/lib/kubelet/pods/7ce3702d-5de0-469b-a8e8-b531c4ea5afa/volumes" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.979794 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95a49175-2149-4cd4-88fc-e25ca5466ff0" path="/var/lib/kubelet/pods/95a49175-2149-4cd4-88fc-e25ca5466ff0/volumes" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.980835 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a5cd6b1-5645-4e0a-be9d-aed6f7930c53" path="/var/lib/kubelet/pods/9a5cd6b1-5645-4e0a-be9d-aed6f7930c53/volumes" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.981455 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c64f8a71-9ee5-4145-b635-bd282dabd781" path="/var/lib/kubelet/pods/c64f8a71-9ee5-4145-b635-bd282dabd781/volumes" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.982039 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e" path="/var/lib/kubelet/pods/c6cd4d13-dc4a-416e-8122-4d4e8c1d5f3e/volumes" Dec 06 07:20:52 crc kubenswrapper[4945]: I1206 07:20:52.983018 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec7d33a0-15af-4bec-b44a-a47263f64043" path="/var/lib/kubelet/pods/ec7d33a0-15af-4bec-b44a-a47263f64043/volumes" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.003340 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bf0fa4a-4a9c-4644-9fe3-413500c59a98" (UID: "0bf0fa4a-4a9c-4644-9fe3-413500c59a98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.024705 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.193893 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae8dae45-f01e-4ef0-8814-07511ff2fcad" (UID: "ae8dae45-f01e-4ef0-8814-07511ff2fcad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.226500 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6d74c1f-98d7-40ac-a1e8-233604971bc9" (UID: "b6d74c1f-98d7-40ac-a1e8-233604971bc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.230833 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.231162 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.273732 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5228457-326b-4470-ad26-f7bba1f50e18" (UID: "d5228457-326b-4470-ad26-f7bba1f50e18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.283726 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efad5b26-d064-4d06-8cc8-361d094dfa9c" (UID: "efad5b26-d064-4d06-8cc8-361d094dfa9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.288814 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-config-data" (OuterVolumeSpecName: "config-data") pod "efad5b26-d064-4d06-8cc8-361d094dfa9c" (UID: "efad5b26-d064-4d06-8cc8-361d094dfa9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.316701 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edbffd82-c356-43a3-94a3-3549b7d9cd31" (UID: "edbffd82-c356-43a3-94a3-3549b7d9cd31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.333389 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.333432 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.333452 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.333463 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: E1206 07:20:53.333610 4945 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 06 07:20:53 crc kubenswrapper[4945]: E1206 07:20:53.333717 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data podName:d71d0084-b4bc-4be5-971b-eced6e6d66ea nodeName:}" failed. No retries permitted until 2025-12-06 07:21:01.333691423 +0000 UTC m=+1674.788552467 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data") pod "rabbitmq-server-0" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea") : configmap "rabbitmq-config-data" not found Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.402057 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.437170 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.470178 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "efad5b26-d064-4d06-8cc8-361d094dfa9c" (UID: "efad5b26-d064-4d06-8cc8-361d094dfa9c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.479157 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "d5228457-326b-4470-ad26-f7bba1f50e18" (UID: "d5228457-326b-4470-ad26-f7bba1f50e18"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.480544 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-config-data" (OuterVolumeSpecName: "config-data") pod "569f0e1c-e58f-4e87-bde3-e7be4659639a" (UID: "569f0e1c-e58f-4e87-bde3-e7be4659639a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.493553 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data" (OuterVolumeSpecName: "config-data") pod "edbffd82-c356-43a3-94a3-3549b7d9cd31" (UID: "edbffd82-c356-43a3-94a3-3549b7d9cd31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.495070 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data" (OuterVolumeSpecName: "config-data") pod "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" (UID: "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.497752 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "ae8dae45-f01e-4ef0-8814-07511ff2fcad" (UID: "ae8dae45-f01e-4ef0-8814-07511ff2fcad"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.502549 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.509728 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "d5228457-326b-4470-ad26-f7bba1f50e18" (UID: "d5228457-326b-4470-ad26-f7bba1f50e18"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.514348 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" (UID: "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.539328 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.539363 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.539374 4945 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.539387 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.539397 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.539409 4945 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/efad5b26-d064-4d06-8cc8-361d094dfa9c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.539419 4945 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d5228457-326b-4470-ad26-f7bba1f50e18-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.539429 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.539437 4945 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae8dae45-f01e-4ef0-8814-07511ff2fcad-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.549389 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" (UID: "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.551940 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "edbffd82-c356-43a3-94a3-3549b7d9cd31" (UID: "edbffd82-c356-43a3-94a3-3549b7d9cd31"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.561462 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "569f0e1c-e58f-4e87-bde3-e7be4659639a" (UID: "569f0e1c-e58f-4e87-bde3-e7be4659639a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.561491 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "569f0e1c-e58f-4e87-bde3-e7be4659639a" (UID: "569f0e1c-e58f-4e87-bde3-e7be4659639a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.573365 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0bf0fa4a-4a9c-4644-9fe3-413500c59a98" (UID: "0bf0fa4a-4a9c-4644-9fe3-413500c59a98"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.574176 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data" (OuterVolumeSpecName: "config-data") pod "b6d74c1f-98d7-40ac-a1e8-233604971bc9" (UID: "b6d74c1f-98d7-40ac-a1e8-233604971bc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.578955 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-config-data" (OuterVolumeSpecName: "config-data") pod "0bf0fa4a-4a9c-4644-9fe3-413500c59a98" (UID: "0bf0fa4a-4a9c-4644-9fe3-413500c59a98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.580430 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" (UID: "5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.584610 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "edbffd82-c356-43a3-94a3-3549b7d9cd31" (UID: "edbffd82-c356-43a3-94a3-3549b7d9cd31"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: E1206 07:20:53.624477 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 06 07:20:53 crc kubenswrapper[4945]: E1206 07:20:53.627446 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.628131 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.631031 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "569f0e1c-e58f-4e87-bde3-e7be4659639a" (UID: "569f0e1c-e58f-4e87-bde3-e7be4659639a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: E1206 07:20:53.631348 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 06 07:20:53 crc kubenswrapper[4945]: E1206 07:20:53.631507 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="232baa8c-2d18-49f3-b230-362181e687df" containerName="ovn-northd" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.640646 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.640679 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.640688 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edbffd82-c356-43a3-94a3-3549b7d9cd31-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.640697 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.640705 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.640713 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.640732 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.640742 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf0fa4a-4a9c-4644-9fe3-413500c59a98-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.640750 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/569f0e1c-e58f-4e87-bde3-e7be4659639a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.640759 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d74c1f-98d7-40ac-a1e8-233604971bc9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.646569 4945 scope.go:117] "RemoveContainer" containerID="1c796787c79166cd1f6bdca3eb60a502cd2b2fc6bd58628b9bff08d4c77ca7c4" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.647658 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.749945 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.750004 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-scripts\") pod \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.750056 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-combined-ca-bundle\") pod \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.750083 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-httpd-run\") pod \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.750118 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-config-data\") pod \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.750137 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-combined-ca-bundle\") pod \"22eb2234-b14b-474c-b1bf-cf726acb064b\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.750161 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s75pv\" (UniqueName: \"kubernetes.io/projected/8ba3e057-d321-48ca-b7a0-8d791cbb7169-kube-api-access-s75pv\") pod \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.750183 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-config-data\") pod \"22eb2234-b14b-474c-b1bf-cf726acb064b\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.750227 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-logs\") pod \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.750335 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt2dc\" (UniqueName: \"kubernetes.io/projected/22eb2234-b14b-474c-b1bf-cf726acb064b-kube-api-access-jt2dc\") pod \"22eb2234-b14b-474c-b1bf-cf726acb064b\" (UID: \"22eb2234-b14b-474c-b1bf-cf726acb064b\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.750362 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-public-tls-certs\") pod \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\" (UID: \"8ba3e057-d321-48ca-b7a0-8d791cbb7169\") " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.758026 4945 scope.go:117] "RemoveContainer" containerID="a7da8e3a50d0a0f1e3ae039655940701da64854727b798ffef58b2a42ba05b35" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.765212 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-logs" (OuterVolumeSpecName: "logs") pod "8ba3e057-d321-48ca-b7a0-8d791cbb7169" (UID: "8ba3e057-d321-48ca-b7a0-8d791cbb7169"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.769101 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8ba3e057-d321-48ca-b7a0-8d791cbb7169" (UID: "8ba3e057-d321-48ca-b7a0-8d791cbb7169"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.781595 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22eb2234-b14b-474c-b1bf-cf726acb064b-kube-api-access-jt2dc" (OuterVolumeSpecName: "kube-api-access-jt2dc") pod "22eb2234-b14b-474c-b1bf-cf726acb064b" (UID: "22eb2234-b14b-474c-b1bf-cf726acb064b"). InnerVolumeSpecName "kube-api-access-jt2dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.790810 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "8ba3e057-d321-48ca-b7a0-8d791cbb7169" (UID: "8ba3e057-d321-48ca-b7a0-8d791cbb7169"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.800098 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-scripts" (OuterVolumeSpecName: "scripts") pod "8ba3e057-d321-48ca-b7a0-8d791cbb7169" (UID: "8ba3e057-d321-48ca-b7a0-8d791cbb7169"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.817696 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystonefa31-account-delete-dt4r7"] Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.823470 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ba3e057-d321-48ca-b7a0-8d791cbb7169-kube-api-access-s75pv" (OuterVolumeSpecName: "kube-api-access-s75pv") pod "8ba3e057-d321-48ca-b7a0-8d791cbb7169" (UID: "8ba3e057-d321-48ca-b7a0-8d791cbb7169"). InnerVolumeSpecName "kube-api-access-s75pv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.834522 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystonefa31-account-delete-dt4r7"] Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.838590 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ba3e057-d321-48ca-b7a0-8d791cbb7169" (UID: "8ba3e057-d321-48ca-b7a0-8d791cbb7169"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.855450 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.855787 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.856354 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.856515 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.856590 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s75pv\" (UniqueName: \"kubernetes.io/projected/8ba3e057-d321-48ca-b7a0-8d791cbb7169-kube-api-access-s75pv\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.856656 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ba3e057-d321-48ca-b7a0-8d791cbb7169-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.856722 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt2dc\" (UniqueName: \"kubernetes.io/projected/22eb2234-b14b-474c-b1bf-cf726acb064b-kube-api-access-jt2dc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.862393 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7db96d494-qst22" event={"ID":"966c9b21-65c4-4d2a-af0f-0041a9bfea08","Type":"ContainerDied","Data":"e5ec440e1b2982d8d5e5efdec9ff243026d7ad0b6f7413eee72b25f1c1bcbb7c"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.862448 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5ec440e1b2982d8d5e5efdec9ff243026d7ad0b6f7413eee72b25f1c1bcbb7c" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.867502 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"22eb2234-b14b-474c-b1bf-cf726acb064b","Type":"ContainerDied","Data":"7f7a222718742cfc4420ef8b9284d3a809008a1d7a2ec4a2c47426e38684a78a"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.867576 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.885365 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-config-data" (OuterVolumeSpecName: "config-data") pod "22eb2234-b14b-474c-b1bf-cf726acb064b" (UID: "22eb2234-b14b-474c-b1bf-cf726acb064b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.908843 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22eb2234-b14b-474c-b1bf-cf726acb064b" (UID: "22eb2234-b14b-474c-b1bf-cf726acb064b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.909920 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron975b-account-delete-nvsg5" event={"ID":"4231a0fc-b0c8-48da-8fab-75cd817dd3ef","Type":"ContainerStarted","Data":"0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.910658 4945 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron975b-account-delete-nvsg5" secret="" err="secret \"galera-openstack-dockercfg-stf7c\" not found" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.915530 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.915700 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ba3e057-d321-48ca-b7a0-8d791cbb7169","Type":"ContainerDied","Data":"e3437617f02eb68845c8942f8f14070ba5170613d8bd7ab6f091ee633fae3a22"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.919471 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9296c3b2-7997-45a0-a7e9-bc1f45ec204e","Type":"ContainerDied","Data":"0c9e7f39a36139d16bd7a520b178089e9c6e969515df2072ac953fdd82cb0235"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.919502 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c9e7f39a36139d16bd7a520b178089e9c6e969515df2072ac953fdd82cb0235" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.921364 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.925990 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0216c-account-delete-bx5zs" event={"ID":"a6fbf826-c8e6-4231-b479-90fe7de5a175","Type":"ContainerStarted","Data":"37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.926118 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell0216c-account-delete-bx5zs" podUID="a6fbf826-c8e6-4231-b479-90fe7de5a175" containerName="mariadb-account-delete" containerID="cri-o://37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da" gracePeriod=30 Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.930449 4945 generic.go:334] "Generic (PLEG): container finished" podID="6ae75a6a-2656-4e65-a84d-8767a570ce8f" containerID="5984506921a2b100cf94ec4a7e72d52586d9d11cdd69b262d6ee89622fc5fcf4" exitCode=0 Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.930507 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ae75a6a-2656-4e65-a84d-8767a570ce8f","Type":"ContainerDied","Data":"5984506921a2b100cf94ec4a7e72d52586d9d11cdd69b262d6ee89622fc5fcf4"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.930531 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ae75a6a-2656-4e65-a84d-8767a570ce8f","Type":"ContainerDied","Data":"4a8573f8c922a12417cbb4b68c50abb4e468a708b2b1f280c89b6ccfbc242cc7"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.930544 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a8573f8c922a12417cbb4b68c50abb4e468a708b2b1f280c89b6ccfbc242cc7" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.932850 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-687fd77bc7-9rms8" event={"ID":"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac","Type":"ContainerDied","Data":"3aaf92a05d038b9022c400636319ff4cd912351991d354de0369dc8c14fb8955"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.932893 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3aaf92a05d038b9022c400636319ff4cd912351991d354de0369dc8c14fb8955" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.938104 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-config-data" (OuterVolumeSpecName: "config-data") pod "8ba3e057-d321-48ca-b7a0-8d791cbb7169" (UID: "8ba3e057-d321-48ca-b7a0-8d791cbb7169"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.943373 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67","Type":"ContainerDied","Data":"66c502aefd40a1a3a51bb03f1f9bedb664bf396043e3ee36b232297da1b7cf8f"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.943445 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66c502aefd40a1a3a51bb03f1f9bedb664bf396043e3ee36b232297da1b7cf8f" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.947440 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8ba3e057-d321-48ca-b7a0-8d791cbb7169" (UID: "8ba3e057-d321-48ca-b7a0-8d791cbb7169"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.947451 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapibadc-account-delete-4n95m" podUID="83eee8a2-f458-4193-8023-ca1228dc92a3" containerName="mariadb-account-delete" containerID="cri-o://a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298" gracePeriod=30 Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.947693 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapibadc-account-delete-4n95m" event={"ID":"83eee8a2-f458-4193-8023-ca1228dc92a3","Type":"ContainerStarted","Data":"a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.950906 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f9668d888-8gtj6" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.952121 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron975b-account-delete-nvsg5" podStartSLOduration=8.952095638 podStartE2EDuration="8.952095638s" podCreationTimestamp="2025-12-06 07:20:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:20:53.922722706 +0000 UTC m=+1667.377583750" watchObservedRunningTime="2025-12-06 07:20:53.952095638 +0000 UTC m=+1667.406956672" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.957999 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.958022 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.958031 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba3e057-d321-48ca-b7a0-8d791cbb7169-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.958040 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.958051 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22eb2234-b14b-474c-b1bf-cf726acb064b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:53 crc kubenswrapper[4945]: E1206 07:20:53.958196 4945 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 06 07:20:53 crc kubenswrapper[4945]: E1206 07:20:53.958264 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts podName:4231a0fc-b0c8-48da-8fab-75cd817dd3ef nodeName:}" failed. No retries permitted until 2025-12-06 07:20:54.458238894 +0000 UTC m=+1667.913099938 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts") pod "neutron975b-account-delete-nvsg5" (UID: "4231a0fc-b0c8-48da-8fab-75cd817dd3ef") : configmap "openstack-scripts" not found Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.965812 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell0216c-account-delete-bx5zs" podStartSLOduration=8.965784748 podStartE2EDuration="8.965784748s" podCreationTimestamp="2025-12-06 07:20:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:20:53.941983316 +0000 UTC m=+1667.396844360" watchObservedRunningTime="2025-12-06 07:20:53.965784748 +0000 UTC m=+1667.420645792" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.990425 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"72607de1-4f23-441e-b722-3844b2eb6d15","Type":"ContainerDied","Data":"7e3e0f97c8029e1212380ce9fe49ecd37ce03778d08cf5c4792094828f4f3476"} Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.990465 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e3e0f97c8029e1212380ce9fe49ecd37ce03778d08cf5c4792094828f4f3476" Dec 06 07:20:53 crc kubenswrapper[4945]: I1206 07:20:53.990842 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapibadc-account-delete-4n95m" podStartSLOduration=7.990810093 podStartE2EDuration="7.990810093s" podCreationTimestamp="2025-12-06 07:20:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:20:53.980903766 +0000 UTC m=+1667.435764820" watchObservedRunningTime="2025-12-06 07:20:53.990810093 +0000 UTC m=+1667.445671137" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.075108 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.080550 4945 scope.go:117] "RemoveContainer" containerID="2388a11738d51f9e160cafe82ed8ca7351af42891b7d72903a3459a847238f95" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.087313 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.121661 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.139916 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.140885 4945 scope.go:117] "RemoveContainer" containerID="f1978a8cc52c8c7435676004e144a0b8373ccf11fd0f8cada610a31d96b4bf20" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.142271 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.163770 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data-custom\") pod \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.163994 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-combined-ca-bundle\") pod \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.164056 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-public-tls-certs\") pod \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.164088 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xckkr\" (UniqueName: \"kubernetes.io/projected/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-kube-api-access-xckkr\") pod \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.164132 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n22t\" (UniqueName: \"kubernetes.io/projected/966c9b21-65c4-4d2a-af0f-0041a9bfea08-kube-api-access-2n22t\") pod \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.164179 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-logs\") pod \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.164227 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-combined-ca-bundle\") pod \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.164267 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data\") pod \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.165382 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-logs" (OuterVolumeSpecName: "logs") pod "9296c3b2-7997-45a0-a7e9-bc1f45ec204e" (UID: "9296c3b2-7997-45a0-a7e9-bc1f45ec204e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.165410 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/966c9b21-65c4-4d2a-af0f-0041a9bfea08-logs\") pod \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\" (UID: \"966c9b21-65c4-4d2a-af0f-0041a9bfea08\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.165521 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-internal-tls-certs\") pod \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.165576 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-config-data\") pod \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\" (UID: \"9296c3b2-7997-45a0-a7e9-bc1f45ec204e\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.165847 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/966c9b21-65c4-4d2a-af0f-0041a9bfea08-logs" (OuterVolumeSpecName: "logs") pod "966c9b21-65c4-4d2a-af0f-0041a9bfea08" (UID: "966c9b21-65c4-4d2a-af0f-0041a9bfea08"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.166922 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.166959 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/966c9b21-65c4-4d2a-af0f-0041a9bfea08-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.174322 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.179925 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-kube-api-access-xckkr" (OuterVolumeSpecName: "kube-api-access-xckkr") pod "9296c3b2-7997-45a0-a7e9-bc1f45ec204e" (UID: "9296c3b2-7997-45a0-a7e9-bc1f45ec204e"). InnerVolumeSpecName "kube-api-access-xckkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.189956 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "966c9b21-65c4-4d2a-af0f-0041a9bfea08" (UID: "966c9b21-65c4-4d2a-af0f-0041a9bfea08"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.198002 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/966c9b21-65c4-4d2a-af0f-0041a9bfea08-kube-api-access-2n22t" (OuterVolumeSpecName: "kube-api-access-2n22t") pod "966c9b21-65c4-4d2a-af0f-0041a9bfea08" (UID: "966c9b21-65c4-4d2a-af0f-0041a9bfea08"). InnerVolumeSpecName "kube-api-access-2n22t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.224948 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-config-data" (OuterVolumeSpecName: "config-data") pod "9296c3b2-7997-45a0-a7e9-bc1f45ec204e" (UID: "9296c3b2-7997-45a0-a7e9-bc1f45ec204e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.236053 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.250128 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9296c3b2-7997-45a0-a7e9-bc1f45ec204e" (UID: "9296c3b2-7997-45a0-a7e9-bc1f45ec204e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.250456 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "966c9b21-65c4-4d2a-af0f-0041a9bfea08" (UID: "966c9b21-65c4-4d2a-af0f-0041a9bfea08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.251191 4945 scope.go:117] "RemoveContainer" containerID="31c19a7223a90ce2817b04d80780886488d735a425cbc88ccf01eeac5899bab6" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268227 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-logs\") pod \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268322 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-combined-ca-bundle\") pod \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268371 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-combined-ca-bundle\") pod \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268400 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp94d\" (UniqueName: \"kubernetes.io/projected/6ae75a6a-2656-4e65-a84d-8767a570ce8f-kube-api-access-vp94d\") pod \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268446 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcmkb\" (UniqueName: \"kubernetes.io/projected/72607de1-4f23-441e-b722-3844b2eb6d15-kube-api-access-mcmkb\") pod \"72607de1-4f23-441e-b722-3844b2eb6d15\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268486 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxbsj\" (UniqueName: \"kubernetes.io/projected/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-kube-api-access-cxbsj\") pod \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268529 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data-custom\") pod \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268603 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-config-data\") pod \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\" (UID: \"6ae75a6a-2656-4e65-a84d-8767a570ce8f\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268771 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-combined-ca-bundle\") pod \"72607de1-4f23-441e-b722-3844b2eb6d15\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268866 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data\") pod \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\" (UID: \"63d2b3b4-2d39-4966-b16c-00e3b9fdbeac\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.268907 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-config-data\") pod \"72607de1-4f23-441e-b722-3844b2eb6d15\" (UID: \"72607de1-4f23-441e-b722-3844b2eb6d15\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.269379 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.269397 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.269412 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.269425 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xckkr\" (UniqueName: \"kubernetes.io/projected/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-kube-api-access-xckkr\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.269437 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n22t\" (UniqueName: \"kubernetes.io/projected/966c9b21-65c4-4d2a-af0f-0041a9bfea08-kube-api-access-2n22t\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.269450 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.275474 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.284984 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9296c3b2-7997-45a0-a7e9-bc1f45ec204e" (UID: "9296c3b2-7997-45a0-a7e9-bc1f45ec204e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.300244 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-logs" (OuterVolumeSpecName: "logs") pod "63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" (UID: "63d2b3b4-2d39-4966-b16c-00e3b9fdbeac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.301663 4945 scope.go:117] "RemoveContainer" containerID="704c83d0821804990d78d0bdd43784abbc4a50001e79e73a668598b93f18f6ac" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.301675 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" (UID: "63d2b3b4-2d39-4966-b16c-00e3b9fdbeac"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.322543 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72607de1-4f23-441e-b722-3844b2eb6d15-kube-api-access-mcmkb" (OuterVolumeSpecName: "kube-api-access-mcmkb") pod "72607de1-4f23-441e-b722-3844b2eb6d15" (UID: "72607de1-4f23-441e-b722-3844b2eb6d15"). InnerVolumeSpecName "kube-api-access-mcmkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.327927 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ae75a6a-2656-4e65-a84d-8767a570ce8f-kube-api-access-vp94d" (OuterVolumeSpecName: "kube-api-access-vp94d") pod "6ae75a6a-2656-4e65-a84d-8767a570ce8f" (UID: "6ae75a6a-2656-4e65-a84d-8767a570ce8f"). InnerVolumeSpecName "kube-api-access-vp94d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.347641 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.356662 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.358739 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-kube-api-access-cxbsj" (OuterVolumeSpecName: "kube-api-access-cxbsj") pod "63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" (UID: "63d2b3b4-2d39-4966-b16c-00e3b9fdbeac"). InnerVolumeSpecName "kube-api-access-cxbsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.363846 4945 scope.go:117] "RemoveContainer" containerID="51e0e29bf34bea465946d1ff2713e5af788a26b0c2039d1356719fdaad5afb73" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.365194 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f9668d888-8gtj6"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.370779 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-memcached-tls-certs\") pod \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.370872 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q8js\" (UniqueName: \"kubernetes.io/projected/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kube-api-access-8q8js\") pod \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.370980 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-config-data\") pod \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.371006 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kolla-config\") pod \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.371040 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-combined-ca-bundle\") pod \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\" (UID: \"ef14eff5-a3d8-48a8-b3fc-b59b97f68b67\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.371478 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.371494 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-logs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.371507 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp94d\" (UniqueName: \"kubernetes.io/projected/6ae75a6a-2656-4e65-a84d-8767a570ce8f-kube-api-access-vp94d\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.371521 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcmkb\" (UniqueName: \"kubernetes.io/projected/72607de1-4f23-441e-b722-3844b2eb6d15-kube-api-access-mcmkb\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.371533 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxbsj\" (UniqueName: \"kubernetes.io/projected/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-kube-api-access-cxbsj\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.371543 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.373482 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-config-data" (OuterVolumeSpecName: "config-data") pod "ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" (UID: "ef14eff5-a3d8-48a8-b3fc-b59b97f68b67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.376258 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5f9668d888-8gtj6"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.381146 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" (UID: "ef14eff5-a3d8-48a8-b3fc-b59b97f68b67"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.385198 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.394049 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kube-api-access-8q8js" (OuterVolumeSpecName: "kube-api-access-8q8js") pod "ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" (UID: "ef14eff5-a3d8-48a8-b3fc-b59b97f68b67"). InnerVolumeSpecName "kube-api-access-8q8js". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.394128 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.401821 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-config-data" (OuterVolumeSpecName: "config-data") pod "72607de1-4f23-441e-b722-3844b2eb6d15" (UID: "72607de1-4f23-441e-b722-3844b2eb6d15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.404101 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.412192 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9296c3b2-7997-45a0-a7e9-bc1f45ec204e" (UID: "9296c3b2-7997-45a0-a7e9-bc1f45ec204e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.412426 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.420080 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.429162 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.441754 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72607de1-4f23-441e-b722-3844b2eb6d15" (UID: "72607de1-4f23-441e-b722-3844b2eb6d15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.451167 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.456402 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.459515 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" (UID: "63d2b3b4-2d39-4966-b16c-00e3b9fdbeac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.459599 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-config-data" (OuterVolumeSpecName: "config-data") pod "6ae75a6a-2656-4e65-a84d-8767a570ce8f" (UID: "6ae75a6a-2656-4e65-a84d-8767a570ce8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.463199 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ae75a6a-2656-4e65-a84d-8767a570ce8f" (UID: "6ae75a6a-2656-4e65-a84d-8767a570ce8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.471347 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.472864 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.472986 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72607de1-4f23-441e-b722-3844b2eb6d15-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.473076 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q8js\" (UniqueName: \"kubernetes.io/projected/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kube-api-access-8q8js\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.473146 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.473209 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.473275 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.473363 4945 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.473440 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae75a6a-2656-4e65-a84d-8767a570ce8f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.473731 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9296c3b2-7997-45a0-a7e9-bc1f45ec204e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: E1206 07:20:54.473862 4945 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 06 07:20:54 crc kubenswrapper[4945]: E1206 07:20:54.473977 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts podName:4231a0fc-b0c8-48da-8fab-75cd817dd3ef nodeName:}" failed. No retries permitted until 2025-12-06 07:20:55.473954869 +0000 UTC m=+1668.928815923 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts") pod "neutron975b-account-delete-nvsg5" (UID: "4231a0fc-b0c8-48da-8fab-75cd817dd3ef") : configmap "openstack-scripts" not found Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.484459 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.501294 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data" (OuterVolumeSpecName: "config-data") pod "966c9b21-65c4-4d2a-af0f-0041a9bfea08" (UID: "966c9b21-65c4-4d2a-af0f-0041a9bfea08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.508118 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-75599744c6-dmqzn"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.520703 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-75599744c6-dmqzn"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.520720 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" (UID: "ef14eff5-a3d8-48a8-b3fc-b59b97f68b67"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.526763 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.534718 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.535413 4945 scope.go:117] "RemoveContainer" containerID="f75d79cb35ec6aeed2fbb5df193689624fa68ec6f36d6f07600634a6a1eb3d4f" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.550625 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data" (OuterVolumeSpecName: "config-data") pod "63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" (UID: "63d2b3b4-2d39-4966-b16c-00e3b9fdbeac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.575514 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.575554 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/966c9b21-65c4-4d2a-af0f-0041a9bfea08-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.575563 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.589754 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" (UID: "ef14eff5-a3d8-48a8-b3fc-b59b97f68b67"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.615467 4945 scope.go:117] "RemoveContainer" containerID="41457071a0e5ace0cbbf96b7adf5ba3e23ee9463d1013d0f9d4b4f9916e1e917" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.671649 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_232baa8c-2d18-49f3-b230-362181e687df/ovn-northd/0.log" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.671751 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.696724 4945 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.696784 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-rbrz6"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.705326 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-rbrz6"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.735093 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder3788-account-delete-56qf8"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.745033 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder3788-account-delete-56qf8"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.751622 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3788-account-create-update-bsdpx"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.758555 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3788-account-create-update-bsdpx"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.798640 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-ovn-northd-tls-certs\") pod \"232baa8c-2d18-49f3-b230-362181e687df\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.798748 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/232baa8c-2d18-49f3-b230-362181e687df-ovn-rundir\") pod \"232baa8c-2d18-49f3-b230-362181e687df\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.798833 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnrx2\" (UniqueName: \"kubernetes.io/projected/232baa8c-2d18-49f3-b230-362181e687df-kube-api-access-lnrx2\") pod \"232baa8c-2d18-49f3-b230-362181e687df\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.798911 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-scripts\") pod \"232baa8c-2d18-49f3-b230-362181e687df\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.798968 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-combined-ca-bundle\") pod \"232baa8c-2d18-49f3-b230-362181e687df\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.799033 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-metrics-certs-tls-certs\") pod \"232baa8c-2d18-49f3-b230-362181e687df\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.799067 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-config\") pod \"232baa8c-2d18-49f3-b230-362181e687df\" (UID: \"232baa8c-2d18-49f3-b230-362181e687df\") " Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.800809 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/232baa8c-2d18-49f3-b230-362181e687df-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "232baa8c-2d18-49f3-b230-362181e687df" (UID: "232baa8c-2d18-49f3-b230-362181e687df"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.801514 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/232baa8c-2d18-49f3-b230-362181e687df-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.802094 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-config" (OuterVolumeSpecName: "config") pod "232baa8c-2d18-49f3-b230-362181e687df" (UID: "232baa8c-2d18-49f3-b230-362181e687df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.804494 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-scripts" (OuterVolumeSpecName: "scripts") pod "232baa8c-2d18-49f3-b230-362181e687df" (UID: "232baa8c-2d18-49f3-b230-362181e687df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.827947 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/232baa8c-2d18-49f3-b230-362181e687df-kube-api-access-lnrx2" (OuterVolumeSpecName: "kube-api-access-lnrx2") pod "232baa8c-2d18-49f3-b230-362181e687df" (UID: "232baa8c-2d18-49f3-b230-362181e687df"). InnerVolumeSpecName "kube-api-access-lnrx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.832889 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-5zkfl"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.846498 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-5zkfl"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.856720 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f244-account-create-update-gvnlc"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.864615 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-f244-account-create-update-gvnlc"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.894624 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancef244-account-delete-vtwn6"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.900807 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "232baa8c-2d18-49f3-b230-362181e687df" (UID: "232baa8c-2d18-49f3-b230-362181e687df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.903072 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.903103 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.903120 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232baa8c-2d18-49f3-b230-362181e687df-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.903131 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnrx2\" (UniqueName: \"kubernetes.io/projected/232baa8c-2d18-49f3-b230-362181e687df-kube-api-access-lnrx2\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.929290 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glancef244-account-delete-vtwn6"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.952084 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-5d2dk"] Dec 06 07:20:54 crc kubenswrapper[4945]: I1206 07:20:54.995041 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" path="/var/lib/kubelet/pods/0bf0fa4a-4a9c-4644-9fe3-413500c59a98/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.000572 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22eb2234-b14b-474c-b1bf-cf726acb064b" path="/var/lib/kubelet/pods/22eb2234-b14b-474c-b1bf-cf726acb064b/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.001252 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="569f0e1c-e58f-4e87-bde3-e7be4659639a" path="/var/lib/kubelet/pods/569f0e1c-e58f-4e87-bde3-e7be4659639a/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.004105 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "232baa8c-2d18-49f3-b230-362181e687df" (UID: "232baa8c-2d18-49f3-b230-362181e687df"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.005006 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.008202 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="593549fe-1eb9-4e90-8a87-b2f620959e98" path="/var/lib/kubelet/pods/593549fe-1eb9-4e90-8a87-b2f620959e98/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.017383 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" path="/var/lib/kubelet/pods/5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.022595 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69560226-9548-42f0-8c4a-cf33fed9b1f6" path="/var/lib/kubelet/pods/69560226-9548-42f0-8c4a-cf33fed9b1f6/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.025447 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8035b30e-21e0-488a-bc9d-78d14f0e11a9" path="/var/lib/kubelet/pods/8035b30e-21e0-488a-bc9d-78d14f0e11a9/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.026208 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" path="/var/lib/kubelet/pods/8ba3e057-d321-48ca-b7a0-8d791cbb7169/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.033760 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.037307 4945 generic.go:334] "Generic (PLEG): container finished" podID="fe30cfa7-9746-4771-bac5-29e3a3662aec" containerID="bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8" exitCode=0 Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.041339 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a35c1b0f-f7ed-47ea-b3b8-dc855c43a729" path="/var/lib/kubelet/pods/a35c1b0f-f7ed-47ea-b3b8-dc855c43a729/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.042344 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "232baa8c-2d18-49f3-b230-362181e687df" (UID: "232baa8c-2d18-49f3-b230-362181e687df"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.047168 4945 generic.go:334] "Generic (PLEG): container finished" podID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" containerID="35755191ab63b793ba68c4e84cc61180235cdedf764a176765a67c7090121592" exitCode=0 Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.047451 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.058339 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" path="/var/lib/kubelet/pods/ae8dae45-f01e-4ef0-8814-07511ff2fcad/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.070735 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" path="/var/lib/kubelet/pods/b6d74c1f-98d7-40ac-a1e8-233604971bc9/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.091785 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1cb89b2-7918-457d-a36a-2cec380f3854" path="/var/lib/kubelet/pods/c1cb89b2-7918-457d-a36a-2cec380f3854/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.077565 4945 generic.go:334] "Generic (PLEG): container finished" podID="5af85716-fd7e-45e9-8320-4db8800302a8" containerID="9d35634fe6ffb6779224176c397d8848048afbbeb060d73a9ed0f8871a1cd24b" exitCode=0 Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.102381 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b" path="/var/lib/kubelet/pods/cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106219 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-plugins-conf\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106320 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-combined-ca-bundle\") pod \"fe30cfa7-9746-4771-bac5-29e3a3662aec\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106373 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p69tr\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-kube-api-access-p69tr\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106415 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-operator-scripts\") pod \"fe30cfa7-9746-4771-bac5-29e3a3662aec\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106452 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/250c85e2-e14d-4f93-822c-c9a88352eafe-erlang-cookie-secret\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106505 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-generated\") pod \"fe30cfa7-9746-4771-bac5-29e3a3662aec\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106557 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-kolla-config\") pod \"fe30cfa7-9746-4771-bac5-29e3a3662aec\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106628 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106701 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-tls\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106737 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-galera-tls-certs\") pod \"fe30cfa7-9746-4771-bac5-29e3a3662aec\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106775 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/250c85e2-e14d-4f93-822c-c9a88352eafe-pod-info\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106804 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-server-conf\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106849 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-default\") pod \"fe30cfa7-9746-4771-bac5-29e3a3662aec\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106903 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.106959 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"fe30cfa7-9746-4771-bac5-29e3a3662aec\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.107015 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2twnc\" (UniqueName: \"kubernetes.io/projected/fe30cfa7-9746-4771-bac5-29e3a3662aec-kube-api-access-2twnc\") pod \"fe30cfa7-9746-4771-bac5-29e3a3662aec\" (UID: \"fe30cfa7-9746-4771-bac5-29e3a3662aec\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.107067 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-plugins\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.107106 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-erlang-cookie\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.107143 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-confd\") pod \"250c85e2-e14d-4f93-822c-c9a88352eafe\" (UID: \"250c85e2-e14d-4f93-822c-c9a88352eafe\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.107242 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.108467 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "fe30cfa7-9746-4771-bac5-29e3a3662aec" (UID: "fe30cfa7-9746-4771-bac5-29e3a3662aec"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.109358 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "fe30cfa7-9746-4771-bac5-29e3a3662aec" (UID: "fe30cfa7-9746-4771-bac5-29e3a3662aec"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.109909 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.109988 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.110514 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "fe30cfa7-9746-4771-bac5-29e3a3662aec" (UID: "fe30cfa7-9746-4771-bac5-29e3a3662aec"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.110566 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5228457-326b-4470-ad26-f7bba1f50e18" path="/var/lib/kubelet/pods/d5228457-326b-4470-ad26-f7bba1f50e18/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.115437 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-kube-api-access-p69tr" (OuterVolumeSpecName: "kube-api-access-p69tr") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "kube-api-access-p69tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117213 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe30cfa7-9746-4771-bac5-29e3a3662aec" (UID: "fe30cfa7-9746-4771-bac5-29e3a3662aec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117665 4945 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117692 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p69tr\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-kube-api-access-p69tr\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117702 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117720 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117729 4945 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117742 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/232baa8c-2d18-49f3-b230-362181e687df-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117750 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe30cfa7-9746-4771-bac5-29e3a3662aec-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117763 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117772 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117801 4945 generic.go:334] "Generic (PLEG): container finished" podID="250c85e2-e14d-4f93-822c-c9a88352eafe" containerID="5a233bae071c35e0cff40c7e5f127d5016758c87b7fe65dcf4b53dce1d196ba5" exitCode=0 Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.117935 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.115222 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.121419 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edbffd82-c356-43a3-94a3-3549b7d9cd31" path="/var/lib/kubelet/pods/edbffd82-c356-43a3-94a3-3549b7d9cd31/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.122268 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" path="/var/lib/kubelet/pods/efad5b26-d064-4d06-8cc8-361d094dfa9c/volumes" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.123571 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.124640 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_232baa8c-2d18-49f3-b230-362181e687df/ovn-northd/0.log" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.124706 4945 generic.go:334] "Generic (PLEG): container finished" podID="232baa8c-2d18-49f3-b230-362181e687df" containerID="01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" exitCode=139 Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.124887 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7db96d494-qst22" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.126189 4945 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron975b-account-delete-nvsg5" secret="" err="secret \"galera-openstack-dockercfg-stf7c\" not found" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.126994 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127346 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fe30cfa7-9746-4771-bac5-29e3a3662aec","Type":"ContainerDied","Data":"bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8"} Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127424 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fe30cfa7-9746-4771-bac5-29e3a3662aec","Type":"ContainerDied","Data":"2a0b523809cf4ba71213abf525a2e8725ce5614df35fd12c49111e1d3470477c"} Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127442 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-5d2dk"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127504 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d71d0084-b4bc-4be5-971b-eced6e6d66ea","Type":"ContainerDied","Data":"35755191ab63b793ba68c4e84cc61180235cdedf764a176765a67c7090121592"} Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127521 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican4ffa-account-delete-r98d4"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127556 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican4ffa-account-delete-r98d4"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127580 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-4ffa-account-create-update-wqkng"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127593 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-4ffa-account-create-update-wqkng"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127607 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-ljfg7"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127623 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-694b7fd487-99w9l" event={"ID":"5af85716-fd7e-45e9-8320-4db8800302a8","Type":"ContainerDied","Data":"9d35634fe6ffb6779224176c397d8848048afbbeb060d73a9ed0f8871a1cd24b"} Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127639 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-ljfg7"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127658 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement73b1-account-delete-l54br"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127672 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"250c85e2-e14d-4f93-822c-c9a88352eafe","Type":"ContainerDied","Data":"5a233bae071c35e0cff40c7e5f127d5016758c87b7fe65dcf4b53dce1d196ba5"} Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127685 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"232baa8c-2d18-49f3-b230-362181e687df","Type":"ContainerDied","Data":"01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4"} Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127697 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"232baa8c-2d18-49f3-b230-362181e687df","Type":"ContainerDied","Data":"0d2515fb2c0ed415c60853fb62ac0440c1b45493e80440f481ade0b520891668"} Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.127728 4945 scope.go:117] "RemoveContainer" containerID="bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.128681 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.129150 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/250c85e2-e14d-4f93-822c-c9a88352eafe-pod-info" (OuterVolumeSpecName: "pod-info") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.129422 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.129510 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe30cfa7-9746-4771-bac5-29e3a3662aec-kube-api-access-2twnc" (OuterVolumeSpecName: "kube-api-access-2twnc") pod "fe30cfa7-9746-4771-bac5-29e3a3662aec" (UID: "fe30cfa7-9746-4771-bac5-29e3a3662aec"). InnerVolumeSpecName "kube-api-access-2twnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.129932 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/250c85e2-e14d-4f93-822c-c9a88352eafe-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.131300 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-687fd77bc7-9rms8" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.132588 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.132681 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.156157 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement73b1-account-delete-l54br"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.166326 4945 scope.go:117] "RemoveContainer" containerID="9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.172232 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "mysql-db") pod "fe30cfa7-9746-4771-bac5-29e3a3662aec" (UID: "fe30cfa7-9746-4771-bac5-29e3a3662aec"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.172507 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-73b1-account-create-update-8sddg"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.179667 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-server-conf" (OuterVolumeSpecName: "server-conf") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.188107 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-73b1-account-create-update-8sddg"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.215512 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe30cfa7-9746-4771-bac5-29e3a3662aec" (UID: "fe30cfa7-9746-4771-bac5-29e3a3662aec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.216222 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data" (OuterVolumeSpecName: "config-data") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.224145 4945 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/250c85e2-e14d-4f93-822c-c9a88352eafe-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.224310 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.224324 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.224333 4945 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.224345 4945 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/250c85e2-e14d-4f93-822c-c9a88352eafe-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.224354 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/250c85e2-e14d-4f93-822c-c9a88352eafe-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.224393 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.224403 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2twnc\" (UniqueName: \"kubernetes.io/projected/fe30cfa7-9746-4771-bac5-29e3a3662aec-kube-api-access-2twnc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.224416 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.248773 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.249700 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.252430 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 07:20:55 crc kubenswrapper[4945]: E1206 07:20:55.252865 4945 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 06 07:20:55 crc kubenswrapper[4945]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-06T07:20:47Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 06 07:20:55 crc kubenswrapper[4945]: /etc/init.d/functions: line 589: 470 Alarm clock "$@" Dec 06 07:20:55 crc kubenswrapper[4945]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-xc2dd" message=< Dec 06 07:20:55 crc kubenswrapper[4945]: Exiting ovn-controller (1) [FAILED] Dec 06 07:20:55 crc kubenswrapper[4945]: Killing ovn-controller (1) [ OK ] Dec 06 07:20:55 crc kubenswrapper[4945]: Killing ovn-controller (1) with SIGKILL [ OK ] Dec 06 07:20:55 crc kubenswrapper[4945]: 2025-12-06T07:20:47Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 06 07:20:55 crc kubenswrapper[4945]: /etc/init.d/functions: line 589: 470 Alarm clock "$@" Dec 06 07:20:55 crc kubenswrapper[4945]: > Dec 06 07:20:55 crc kubenswrapper[4945]: E1206 07:20:55.252889 4945 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 06 07:20:55 crc kubenswrapper[4945]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-06T07:20:47Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 06 07:20:55 crc kubenswrapper[4945]: /etc/init.d/functions: line 589: 470 Alarm clock "$@" Dec 06 07:20:55 crc kubenswrapper[4945]: > pod="openstack/ovn-controller-xc2dd" podUID="3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" containerName="ovn-controller" containerID="cri-o://40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.252954 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-xc2dd" podUID="3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" containerName="ovn-controller" containerID="cri-o://40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398" gracePeriod=22 Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.259382 4945 scope.go:117] "RemoveContainer" containerID="bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8" Dec 06 07:20:55 crc kubenswrapper[4945]: E1206 07:20:55.259941 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8\": container with ID starting with bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8 not found: ID does not exist" containerID="bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.259967 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8"} err="failed to get container status \"bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8\": rpc error: code = NotFound desc = could not find container \"bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8\": container with ID starting with bf5648c427c71f4ef11c5b80feed25022a774ee650f980791bab8b65cbaf50d8 not found: ID does not exist" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.260005 4945 scope.go:117] "RemoveContainer" containerID="9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b" Dec 06 07:20:55 crc kubenswrapper[4945]: E1206 07:20:55.260419 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b\": container with ID starting with 9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b not found: ID does not exist" containerID="9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.260466 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b"} err="failed to get container status \"9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b\": rpc error: code = NotFound desc = could not find container \"9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b\": container with ID starting with 9edcd869c0ff45cf269b11bdd23360629e72c6eff5dd8ab608be6a32ee539e9b not found: ID does not exist" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.260500 4945 scope.go:117] "RemoveContainer" containerID="5a233bae071c35e0cff40c7e5f127d5016758c87b7fe65dcf4b53dce1d196ba5" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.291577 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "fe30cfa7-9746-4771-bac5-29e3a3662aec" (UID: "fe30cfa7-9746-4771-bac5-29e3a3662aec"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.324851 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-server-conf\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.325079 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-plugins\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.325168 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-tls\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.325244 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-plugins-conf\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.325345 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d71d0084-b4bc-4be5-971b-eced6e6d66ea-erlang-cookie-secret\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.325425 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d5hl\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-kube-api-access-6d5hl\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.325550 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.325629 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.325707 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-confd\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.325780 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-erlang-cookie\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.325890 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d71d0084-b4bc-4be5-971b-eced6e6d66ea-pod-info\") pod \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\" (UID: \"d71d0084-b4bc-4be5-971b-eced6e6d66ea\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.327033 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.327067 4945 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe30cfa7-9746-4771-bac5-29e3a3662aec-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.327080 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.328192 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.329246 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.329507 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.330530 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.333178 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.333288 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "250c85e2-e14d-4f93-822c-c9a88352eafe" (UID: "250c85e2-e14d-4f93-822c-c9a88352eafe"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.333377 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d71d0084-b4bc-4be5-971b-eced6e6d66ea-pod-info" (OuterVolumeSpecName: "pod-info") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.333483 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-kube-api-access-6d5hl" (OuterVolumeSpecName: "kube-api-access-6d5hl") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "kube-api-access-6d5hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.336528 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d71d0084-b4bc-4be5-971b-eced6e6d66ea-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.355614 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data" (OuterVolumeSpecName: "config-data") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.379675 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-server-conf" (OuterVolumeSpecName: "server-conf") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.418602 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.425038 4945 scope.go:117] "RemoveContainer" containerID="3da187010c6d527c479f778bf05d2b504b2c0f1fcf3e34a35517eca9354e58fd" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428156 4945 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428192 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428207 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428219 4945 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428230 4945 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d71d0084-b4bc-4be5-971b-eced6e6d66ea-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428242 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d5hl\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-kube-api-access-6d5hl\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428254 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/250c85e2-e14d-4f93-822c-c9a88352eafe-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428306 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428320 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d71d0084-b4bc-4be5-971b-eced6e6d66ea-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428333 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.428347 4945 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d71d0084-b4bc-4be5-971b-eced6e6d66ea-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.431401 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.440131 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.462405 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.467058 4945 scope.go:117] "RemoveContainer" containerID="db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.467215 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d71d0084-b4bc-4be5-971b-eced6e6d66ea" (UID: "d71d0084-b4bc-4be5-971b-eced6e6d66ea"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.491665 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.494435 4945 scope.go:117] "RemoveContainer" containerID="01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.501137 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.529827 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-credential-keys\") pod \"5af85716-fd7e-45e9-8320-4db8800302a8\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.529875 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-combined-ca-bundle\") pod \"5af85716-fd7e-45e9-8320-4db8800302a8\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.529996 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-config-data\") pod \"5af85716-fd7e-45e9-8320-4db8800302a8\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.530019 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-scripts\") pod \"5af85716-fd7e-45e9-8320-4db8800302a8\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.530065 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-public-tls-certs\") pod \"5af85716-fd7e-45e9-8320-4db8800302a8\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.530120 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpwxr\" (UniqueName: \"kubernetes.io/projected/5af85716-fd7e-45e9-8320-4db8800302a8-kube-api-access-fpwxr\") pod \"5af85716-fd7e-45e9-8320-4db8800302a8\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.530165 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-internal-tls-certs\") pod \"5af85716-fd7e-45e9-8320-4db8800302a8\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.530185 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-fernet-keys\") pod \"5af85716-fd7e-45e9-8320-4db8800302a8\" (UID: \"5af85716-fd7e-45e9-8320-4db8800302a8\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.531053 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.531070 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d71d0084-b4bc-4be5-971b-eced6e6d66ea-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: E1206 07:20:55.531140 4945 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 06 07:20:55 crc kubenswrapper[4945]: E1206 07:20:55.531189 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts podName:4231a0fc-b0c8-48da-8fab-75cd817dd3ef nodeName:}" failed. No retries permitted until 2025-12-06 07:20:57.531172594 +0000 UTC m=+1670.986033638 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts") pod "neutron975b-account-delete-nvsg5" (UID: "4231a0fc-b0c8-48da-8fab-75cd817dd3ef") : configmap "openstack-scripts" not found Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.547609 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.550611 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5af85716-fd7e-45e9-8320-4db8800302a8" (UID: "5af85716-fd7e-45e9-8320-4db8800302a8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.550625 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.553429 4945 scope.go:117] "RemoveContainer" containerID="db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.553734 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af85716-fd7e-45e9-8320-4db8800302a8-kube-api-access-fpwxr" (OuterVolumeSpecName: "kube-api-access-fpwxr") pod "5af85716-fd7e-45e9-8320-4db8800302a8" (UID: "5af85716-fd7e-45e9-8320-4db8800302a8"). InnerVolumeSpecName "kube-api-access-fpwxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.559546 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.564813 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5af85716-fd7e-45e9-8320-4db8800302a8" (UID: "5af85716-fd7e-45e9-8320-4db8800302a8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.565666 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: E1206 07:20:55.569874 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba\": container with ID starting with db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba not found: ID does not exist" containerID="db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.569930 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba"} err="failed to get container status \"db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba\": rpc error: code = NotFound desc = could not find container \"db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba\": container with ID starting with db589aba47c38038889d31bb51c18f372e54baca61efc78758feef0c3b0b66ba not found: ID does not exist" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.569960 4945 scope.go:117] "RemoveContainer" containerID="01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" Dec 06 07:20:55 crc kubenswrapper[4945]: E1206 07:20:55.572430 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4\": container with ID starting with 01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4 not found: ID does not exist" containerID="01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.572492 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4"} err="failed to get container status \"01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4\": rpc error: code = NotFound desc = could not find container \"01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4\": container with ID starting with 01c731c622e35314f1768e833788eae4fcea3797fa3b8fe54eb63f85ba47def4 not found: ID does not exist" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.572606 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.574587 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-scripts" (OuterVolumeSpecName: "scripts") pod "5af85716-fd7e-45e9-8320-4db8800302a8" (UID: "5af85716-fd7e-45e9-8320-4db8800302a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.577593 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.584431 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.589538 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.595085 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7db96d494-qst22"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.602835 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-7db96d494-qst22"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.606469 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5af85716-fd7e-45e9-8320-4db8800302a8" (UID: "5af85716-fd7e-45e9-8320-4db8800302a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.606939 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-687fd77bc7-9rms8"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.612026 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-687fd77bc7-9rms8"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.636655 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpwxr\" (UniqueName: \"kubernetes.io/projected/5af85716-fd7e-45e9-8320-4db8800302a8-kube-api-access-fpwxr\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.636697 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.636708 4945 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.636718 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.636728 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.636987 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-config-data" (OuterVolumeSpecName: "config-data") pod "5af85716-fd7e-45e9-8320-4db8800302a8" (UID: "5af85716-fd7e-45e9-8320-4db8800302a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.646888 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-xc2dd_3fb32cca-c8f2-4ff6-a05c-25f3e56b5478/ovn-controller/0.log" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.646980 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xc2dd" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.657452 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5af85716-fd7e-45e9-8320-4db8800302a8" (UID: "5af85716-fd7e-45e9-8320-4db8800302a8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.675032 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5af85716-fd7e-45e9-8320-4db8800302a8" (UID: "5af85716-fd7e-45e9-8320-4db8800302a8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.741425 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-combined-ca-bundle\") pod \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.741570 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run\") pod \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.741669 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run" (OuterVolumeSpecName: "var-run") pod "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" (UID: "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.741714 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-ovn-controller-tls-certs\") pod \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742114 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run-ovn\") pod \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742138 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-scripts\") pod \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742160 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-log-ovn\") pod \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742176 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkrf2\" (UniqueName: \"kubernetes.io/projected/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-kube-api-access-pkrf2\") pod \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\" (UID: \"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478\") " Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742376 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" (UID: "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742509 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" (UID: "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742783 4945 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742809 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742835 4945 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742848 4945 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742858 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.742872 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af85716-fd7e-45e9-8320-4db8800302a8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.752248 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-4vz7h"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.757248 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-scripts" (OuterVolumeSpecName: "scripts") pod "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" (UID: "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.772534 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-kube-api-access-pkrf2" (OuterVolumeSpecName: "kube-api-access-pkrf2") pod "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" (UID: "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478"). InnerVolumeSpecName "kube-api-access-pkrf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.820350 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-4vz7h"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.820359 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" (UID: "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.851242 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.851293 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkrf2\" (UniqueName: \"kubernetes.io/projected/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-kube-api-access-pkrf2\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.851308 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.890324 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-975b-account-create-update-rhkgz"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.969437 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron975b-account-delete-nvsg5"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.973894 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-975b-account-create-update-rhkgz"] Dec 06 07:20:55 crc kubenswrapper[4945]: I1206 07:20:55.986466 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" (UID: "3fb32cca-c8f2-4ff6-a05c-25f3e56b5478"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.058809 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:56 crc kubenswrapper[4945]: E1206 07:20:56.134602 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:20:56 crc kubenswrapper[4945]: E1206 07:20:56.135071 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:20:56 crc kubenswrapper[4945]: E1206 07:20:56.135366 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:20:56 crc kubenswrapper[4945]: E1206 07:20:56.135423 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:20:56 crc kubenswrapper[4945]: E1206 07:20:56.135449 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" Dec 06 07:20:56 crc kubenswrapper[4945]: E1206 07:20:56.137125 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:20:56 crc kubenswrapper[4945]: E1206 07:20:56.138264 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:20:56 crc kubenswrapper[4945]: E1206 07:20:56.138322 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovs-vswitchd" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.148548 4945 generic.go:334] "Generic (PLEG): container finished" podID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerID="b9f8060a9b781ba6b1a59825c6f03e6c1703a2845c803ebc8a4b8e3e321741c0" exitCode=0 Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.148640 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8fbb8a7-4226-4b2f-9e95-621059e6283d","Type":"ContainerDied","Data":"b9f8060a9b781ba6b1a59825c6f03e6c1703a2845c803ebc8a4b8e3e321741c0"} Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.151808 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-694b7fd487-99w9l" event={"ID":"5af85716-fd7e-45e9-8320-4db8800302a8","Type":"ContainerDied","Data":"df625dfa320e111d2f7261b8097860d959f7d506b7125473808d8951af4edb8b"} Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.151869 4945 scope.go:117] "RemoveContainer" containerID="9d35634fe6ffb6779224176c397d8848048afbbeb060d73a9ed0f8871a1cd24b" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.151999 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-694b7fd487-99w9l" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.178732 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.179638 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d71d0084-b4bc-4be5-971b-eced6e6d66ea","Type":"ContainerDied","Data":"44bc0c48bbbaf35075c884c6fb25ca15b8147b4de1fb584d714e11d7ffaf858b"} Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.184899 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.203399 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-694b7fd487-99w9l"] Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.217523 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-xc2dd_3fb32cca-c8f2-4ff6-a05c-25f3e56b5478/ovn-controller/0.log" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.217588 4945 generic.go:334] "Generic (PLEG): container finished" podID="3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" containerID="40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398" exitCode=137 Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.217731 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xc2dd" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.217851 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xc2dd" event={"ID":"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478","Type":"ContainerDied","Data":"40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398"} Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.217888 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xc2dd" event={"ID":"3fb32cca-c8f2-4ff6-a05c-25f3e56b5478","Type":"ContainerDied","Data":"e49cafcdd2705b328ea643d4240d8c3a8df1612a9c54575dfeb9cfa088329fc5"} Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.218058 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron975b-account-delete-nvsg5" podUID="4231a0fc-b0c8-48da-8fab-75cd817dd3ef" containerName="mariadb-account-delete" containerID="cri-o://0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c" gracePeriod=30 Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.248392 4945 scope.go:117] "RemoveContainer" containerID="35755191ab63b793ba68c4e84cc61180235cdedf764a176765a67c7090121592" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.254983 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-694b7fd487-99w9l"] Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.269641 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.279957 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.280426 4945 scope.go:117] "RemoveContainer" containerID="299978e81ebf6ee60eff482d48ee097c771fe6fe34e8dc8d956786247712867f" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.298141 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.309123 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.316039 4945 scope.go:117] "RemoveContainer" containerID="40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.320642 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xc2dd"] Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.338019 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-xc2dd"] Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.340499 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.353694 4945 scope.go:117] "RemoveContainer" containerID="40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398" Dec 06 07:20:56 crc kubenswrapper[4945]: E1206 07:20:56.359971 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398\": container with ID starting with 40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398 not found: ID does not exist" containerID="40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.360037 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398"} err="failed to get container status \"40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398\": rpc error: code = NotFound desc = could not find container \"40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398\": container with ID starting with 40a95d13b412feed7b210fd5de3d9e13655128cbd26f32b3c09854cb25266398 not found: ID does not exist" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.480554 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-log-httpd\") pod \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.480646 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-config-data\") pod \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.480666 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-combined-ca-bundle\") pod \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.480735 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5wqh\" (UniqueName: \"kubernetes.io/projected/f8fbb8a7-4226-4b2f-9e95-621059e6283d-kube-api-access-d5wqh\") pod \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.480789 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-sg-core-conf-yaml\") pod \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.480804 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-scripts\") pod \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.480841 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-ceilometer-tls-certs\") pod \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.480874 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-run-httpd\") pod \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\" (UID: \"f8fbb8a7-4226-4b2f-9e95-621059e6283d\") " Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.481311 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f8fbb8a7-4226-4b2f-9e95-621059e6283d" (UID: "f8fbb8a7-4226-4b2f-9e95-621059e6283d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.481487 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f8fbb8a7-4226-4b2f-9e95-621059e6283d" (UID: "f8fbb8a7-4226-4b2f-9e95-621059e6283d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.486557 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8fbb8a7-4226-4b2f-9e95-621059e6283d-kube-api-access-d5wqh" (OuterVolumeSpecName: "kube-api-access-d5wqh") pod "f8fbb8a7-4226-4b2f-9e95-621059e6283d" (UID: "f8fbb8a7-4226-4b2f-9e95-621059e6283d"). InnerVolumeSpecName "kube-api-access-d5wqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.487386 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-scripts" (OuterVolumeSpecName: "scripts") pod "f8fbb8a7-4226-4b2f-9e95-621059e6283d" (UID: "f8fbb8a7-4226-4b2f-9e95-621059e6283d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.514408 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f8fbb8a7-4226-4b2f-9e95-621059e6283d" (UID: "f8fbb8a7-4226-4b2f-9e95-621059e6283d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.544483 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f8fbb8a7-4226-4b2f-9e95-621059e6283d" (UID: "f8fbb8a7-4226-4b2f-9e95-621059e6283d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.555959 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8fbb8a7-4226-4b2f-9e95-621059e6283d" (UID: "f8fbb8a7-4226-4b2f-9e95-621059e6283d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.593393 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.593435 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.593447 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.593458 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.593469 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8fbb8a7-4226-4b2f-9e95-621059e6283d-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.593478 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.593492 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5wqh\" (UniqueName: \"kubernetes.io/projected/f8fbb8a7-4226-4b2f-9e95-621059e6283d-kube-api-access-d5wqh\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.608495 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-config-data" (OuterVolumeSpecName: "config-data") pod "f8fbb8a7-4226-4b2f-9e95-621059e6283d" (UID: "f8fbb8a7-4226-4b2f-9e95-621059e6283d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.702496 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8fbb8a7-4226-4b2f-9e95-621059e6283d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.794792 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="d5228457-326b-4470-ad26-f7bba1f50e18" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.180:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.963918 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10ba62ff-07ab-489e-84b6-059d2299e606" path="/var/lib/kubelet/pods/10ba62ff-07ab-489e-84b6-059d2299e606/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.964776 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="232baa8c-2d18-49f3-b230-362181e687df" path="/var/lib/kubelet/pods/232baa8c-2d18-49f3-b230-362181e687df/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.965612 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="250c85e2-e14d-4f93-822c-c9a88352eafe" path="/var/lib/kubelet/pods/250c85e2-e14d-4f93-822c-c9a88352eafe/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.966721 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25f2d927-0251-437d-a331-4a962f28f7bc" path="/var/lib/kubelet/pods/25f2d927-0251-437d-a331-4a962f28f7bc/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.967250 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35daefb6-843e-4791-8ef8-c9f850b7bf43" path="/var/lib/kubelet/pods/35daefb6-843e-4791-8ef8-c9f850b7bf43/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.967873 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" path="/var/lib/kubelet/pods/3fb32cca-c8f2-4ff6-a05c-25f3e56b5478/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.968965 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af85716-fd7e-45e9-8320-4db8800302a8" path="/var/lib/kubelet/pods/5af85716-fd7e-45e9-8320-4db8800302a8/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.969472 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e3c48ac-c191-4478-877e-9600b6f63929" path="/var/lib/kubelet/pods/5e3c48ac-c191-4478-877e-9600b6f63929/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.970024 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" path="/var/lib/kubelet/pods/63d2b3b4-2d39-4966-b16c-00e3b9fdbeac/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.971214 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ae75a6a-2656-4e65-a84d-8767a570ce8f" path="/var/lib/kubelet/pods/6ae75a6a-2656-4e65-a84d-8767a570ce8f/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.971715 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72607de1-4f23-441e-b722-3844b2eb6d15" path="/var/lib/kubelet/pods/72607de1-4f23-441e-b722-3844b2eb6d15/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.972293 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="824208f4-ec79-494c-9340-b14d7c7c78b6" path="/var/lib/kubelet/pods/824208f4-ec79-494c-9340-b14d7c7c78b6/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.973399 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" path="/var/lib/kubelet/pods/9296c3b2-7997-45a0-a7e9-bc1f45ec204e/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.973958 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95dc303d-9736-4c29-a1f4-f8b362179eeb" path="/var/lib/kubelet/pods/95dc303d-9736-4c29-a1f4-f8b362179eeb/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.978469 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" path="/var/lib/kubelet/pods/966c9b21-65c4-4d2a-af0f-0041a9bfea08/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.979545 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" path="/var/lib/kubelet/pods/d71d0084-b4bc-4be5-971b-eced6e6d66ea/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.980166 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd562125-d6e3-4f47-8bc0-cace5b235a27" path="/var/lib/kubelet/pods/dd562125-d6e3-4f47-8bc0-cace5b235a27/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.981381 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e61d88c2-c6e2-428e-985d-4d49948f5c0f" path="/var/lib/kubelet/pods/e61d88c2-c6e2-428e-985d-4d49948f5c0f/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.981935 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" path="/var/lib/kubelet/pods/ef14eff5-a3d8-48a8-b3fc-b59b97f68b67/volumes" Dec 06 07:20:56 crc kubenswrapper[4945]: I1206 07:20:56.982644 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe30cfa7-9746-4771-bac5-29e3a3662aec" path="/var/lib/kubelet/pods/fe30cfa7-9746-4771-bac5-29e3a3662aec/volumes" Dec 06 07:20:57 crc kubenswrapper[4945]: I1206 07:20:57.233375 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8fbb8a7-4226-4b2f-9e95-621059e6283d","Type":"ContainerDied","Data":"d039eb69ff87ec7d06020cd72504dffba2a65474ffb2860d2ec9afed1e396362"} Dec 06 07:20:57 crc kubenswrapper[4945]: I1206 07:20:57.233449 4945 scope.go:117] "RemoveContainer" containerID="adc583ba09127799a295f5c8a757b2e8cd330c17f41b39958b390affbc31d557" Dec 06 07:20:57 crc kubenswrapper[4945]: I1206 07:20:57.233491 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:20:57 crc kubenswrapper[4945]: I1206 07:20:57.258803 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:20:57 crc kubenswrapper[4945]: I1206 07:20:57.264512 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:20:57 crc kubenswrapper[4945]: I1206 07:20:57.265517 4945 scope.go:117] "RemoveContainer" containerID="9e0f28c738c6aecbb802ee0805c8bd63678cb98319f526693685a2598a4babc5" Dec 06 07:20:57 crc kubenswrapper[4945]: I1206 07:20:57.284982 4945 scope.go:117] "RemoveContainer" containerID="b9f8060a9b781ba6b1a59825c6f03e6c1703a2845c803ebc8a4b8e3e321741c0" Dec 06 07:20:57 crc kubenswrapper[4945]: I1206 07:20:57.307595 4945 scope.go:117] "RemoveContainer" containerID="18675bd130adeb9a967975c3bf8070d18d7c2f22354b8443abed915b5b687156" Dec 06 07:20:57 crc kubenswrapper[4945]: E1206 07:20:57.617453 4945 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 06 07:20:57 crc kubenswrapper[4945]: E1206 07:20:57.617566 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts podName:4231a0fc-b0c8-48da-8fab-75cd817dd3ef nodeName:}" failed. No retries permitted until 2025-12-06 07:21:01.617540438 +0000 UTC m=+1675.072401482 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts") pod "neutron975b-account-delete-nvsg5" (UID: "4231a0fc-b0c8-48da-8fab-75cd817dd3ef") : configmap "openstack-scripts" not found Dec 06 07:20:58 crc kubenswrapper[4945]: I1206 07:20:58.967923 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" path="/var/lib/kubelet/pods/f8fbb8a7-4226-4b2f-9e95-621059e6283d/volumes" Dec 06 07:21:01 crc kubenswrapper[4945]: E1206 07:21:01.133310 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:01 crc kubenswrapper[4945]: E1206 07:21:01.134523 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:01 crc kubenswrapper[4945]: E1206 07:21:01.134855 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:01 crc kubenswrapper[4945]: E1206 07:21:01.134902 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" Dec 06 07:21:01 crc kubenswrapper[4945]: E1206 07:21:01.135568 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:01 crc kubenswrapper[4945]: E1206 07:21:01.136917 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:01 crc kubenswrapper[4945]: E1206 07:21:01.138532 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:01 crc kubenswrapper[4945]: E1206 07:21:01.138579 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovs-vswitchd" Dec 06 07:21:01 crc kubenswrapper[4945]: E1206 07:21:01.679820 4945 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 06 07:21:01 crc kubenswrapper[4945]: E1206 07:21:01.679930 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts podName:4231a0fc-b0c8-48da-8fab-75cd817dd3ef nodeName:}" failed. No retries permitted until 2025-12-06 07:21:09.679902378 +0000 UTC m=+1683.134763412 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts") pod "neutron975b-account-delete-nvsg5" (UID: "4231a0fc-b0c8-48da-8fab-75cd817dd3ef") : configmap "openstack-scripts" not found Dec 06 07:21:04 crc kubenswrapper[4945]: I1206 07:21:04.942048 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7b595c49d9-jngxp" podUID="85ab2778-4811-4700-bca0-d55c9282699f" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.152:9696/\": dial tcp 10.217.0.152:9696: connect: connection refused" Dec 06 07:21:06 crc kubenswrapper[4945]: E1206 07:21:06.133559 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:06 crc kubenswrapper[4945]: E1206 07:21:06.134491 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:06 crc kubenswrapper[4945]: E1206 07:21:06.134701 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:06 crc kubenswrapper[4945]: E1206 07:21:06.135082 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:06 crc kubenswrapper[4945]: E1206 07:21:06.135115 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" Dec 06 07:21:06 crc kubenswrapper[4945]: E1206 07:21:06.135785 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:06 crc kubenswrapper[4945]: E1206 07:21:06.136802 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:06 crc kubenswrapper[4945]: E1206 07:21:06.136841 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovs-vswitchd" Dec 06 07:21:08 crc kubenswrapper[4945]: I1206 07:21:08.796385 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:21:08 crc kubenswrapper[4945]: I1206 07:21:08.796816 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:21:08 crc kubenswrapper[4945]: I1206 07:21:08.796869 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:21:08 crc kubenswrapper[4945]: I1206 07:21:08.797585 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:21:08 crc kubenswrapper[4945]: I1206 07:21:08.797639 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" gracePeriod=600 Dec 06 07:21:08 crc kubenswrapper[4945]: E1206 07:21:08.933512 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:21:09 crc kubenswrapper[4945]: I1206 07:21:09.366496 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" exitCode=0 Dec 06 07:21:09 crc kubenswrapper[4945]: I1206 07:21:09.366534 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed"} Dec 06 07:21:09 crc kubenswrapper[4945]: I1206 07:21:09.366564 4945 scope.go:117] "RemoveContainer" containerID="6db8402445675a35c1cfea5f14ca674ed295b8ef4463d76bc429b4cab4e4331a" Dec 06 07:21:09 crc kubenswrapper[4945]: I1206 07:21:09.367256 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:21:09 crc kubenswrapper[4945]: E1206 07:21:09.367579 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:21:09 crc kubenswrapper[4945]: E1206 07:21:09.746509 4945 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 06 07:21:09 crc kubenswrapper[4945]: E1206 07:21:09.746579 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts podName:4231a0fc-b0c8-48da-8fab-75cd817dd3ef nodeName:}" failed. No retries permitted until 2025-12-06 07:21:25.746562019 +0000 UTC m=+1699.201423063 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts") pod "neutron975b-account-delete-nvsg5" (UID: "4231a0fc-b0c8-48da-8fab-75cd817dd3ef") : configmap "openstack-scripts" not found Dec 06 07:21:09 crc kubenswrapper[4945]: I1206 07:21:09.916207 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.049992 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf8sb\" (UniqueName: \"kubernetes.io/projected/85ab2778-4811-4700-bca0-d55c9282699f-kube-api-access-kf8sb\") pod \"85ab2778-4811-4700-bca0-d55c9282699f\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.050052 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-public-tls-certs\") pod \"85ab2778-4811-4700-bca0-d55c9282699f\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.050135 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-ovndb-tls-certs\") pod \"85ab2778-4811-4700-bca0-d55c9282699f\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.050169 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-config\") pod \"85ab2778-4811-4700-bca0-d55c9282699f\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.050206 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-internal-tls-certs\") pod \"85ab2778-4811-4700-bca0-d55c9282699f\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.050227 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-combined-ca-bundle\") pod \"85ab2778-4811-4700-bca0-d55c9282699f\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.050296 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-httpd-config\") pod \"85ab2778-4811-4700-bca0-d55c9282699f\" (UID: \"85ab2778-4811-4700-bca0-d55c9282699f\") " Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.056636 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85ab2778-4811-4700-bca0-d55c9282699f-kube-api-access-kf8sb" (OuterVolumeSpecName: "kube-api-access-kf8sb") pod "85ab2778-4811-4700-bca0-d55c9282699f" (UID: "85ab2778-4811-4700-bca0-d55c9282699f"). InnerVolumeSpecName "kube-api-access-kf8sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.056733 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "85ab2778-4811-4700-bca0-d55c9282699f" (UID: "85ab2778-4811-4700-bca0-d55c9282699f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.095665 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85ab2778-4811-4700-bca0-d55c9282699f" (UID: "85ab2778-4811-4700-bca0-d55c9282699f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.097149 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-config" (OuterVolumeSpecName: "config") pod "85ab2778-4811-4700-bca0-d55c9282699f" (UID: "85ab2778-4811-4700-bca0-d55c9282699f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.099121 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "85ab2778-4811-4700-bca0-d55c9282699f" (UID: "85ab2778-4811-4700-bca0-d55c9282699f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.104267 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "85ab2778-4811-4700-bca0-d55c9282699f" (UID: "85ab2778-4811-4700-bca0-d55c9282699f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.111974 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "85ab2778-4811-4700-bca0-d55c9282699f" (UID: "85ab2778-4811-4700-bca0-d55c9282699f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.152107 4945 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.152148 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.152158 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.152169 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.152177 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.152186 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf8sb\" (UniqueName: \"kubernetes.io/projected/85ab2778-4811-4700-bca0-d55c9282699f-kube-api-access-kf8sb\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.152197 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ab2778-4811-4700-bca0-d55c9282699f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.379942 4945 generic.go:334] "Generic (PLEG): container finished" podID="85ab2778-4811-4700-bca0-d55c9282699f" containerID="915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea" exitCode=0 Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.380005 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b595c49d9-jngxp" event={"ID":"85ab2778-4811-4700-bca0-d55c9282699f","Type":"ContainerDied","Data":"915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea"} Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.380032 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b595c49d9-jngxp" event={"ID":"85ab2778-4811-4700-bca0-d55c9282699f","Type":"ContainerDied","Data":"29d2871cbbd51f2c03441cdf4a01555e9f6c5623844918cfeecdd50cd701d0b8"} Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.380049 4945 scope.go:117] "RemoveContainer" containerID="9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.380132 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b595c49d9-jngxp" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.410432 4945 scope.go:117] "RemoveContainer" containerID="915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.417303 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7b595c49d9-jngxp"] Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.425446 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7b595c49d9-jngxp"] Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.443226 4945 scope.go:117] "RemoveContainer" containerID="9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6" Dec 06 07:21:10 crc kubenswrapper[4945]: E1206 07:21:10.443749 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6\": container with ID starting with 9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6 not found: ID does not exist" containerID="9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.443779 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6"} err="failed to get container status \"9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6\": rpc error: code = NotFound desc = could not find container \"9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6\": container with ID starting with 9564f714000902380b0b303e93b2cc2ce58cc196ca246837809dd59d30a583c6 not found: ID does not exist" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.443799 4945 scope.go:117] "RemoveContainer" containerID="915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea" Dec 06 07:21:10 crc kubenswrapper[4945]: E1206 07:21:10.444193 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea\": container with ID starting with 915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea not found: ID does not exist" containerID="915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.444234 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea"} err="failed to get container status \"915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea\": rpc error: code = NotFound desc = could not find container \"915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea\": container with ID starting with 915cf998a55c215799c259ce5fba2ffc445ec73c02fdd63043c619d624227eea not found: ID does not exist" Dec 06 07:21:10 crc kubenswrapper[4945]: I1206 07:21:10.969058 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85ab2778-4811-4700-bca0-d55c9282699f" path="/var/lib/kubelet/pods/85ab2778-4811-4700-bca0-d55c9282699f/volumes" Dec 06 07:21:11 crc kubenswrapper[4945]: E1206 07:21:11.133548 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:11 crc kubenswrapper[4945]: E1206 07:21:11.134164 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:11 crc kubenswrapper[4945]: E1206 07:21:11.134778 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:11 crc kubenswrapper[4945]: E1206 07:21:11.134874 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" Dec 06 07:21:11 crc kubenswrapper[4945]: E1206 07:21:11.136937 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:11 crc kubenswrapper[4945]: E1206 07:21:11.138474 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:11 crc kubenswrapper[4945]: E1206 07:21:11.139900 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:11 crc kubenswrapper[4945]: E1206 07:21:11.139942 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovs-vswitchd" Dec 06 07:21:13 crc kubenswrapper[4945]: E1206 07:21:13.829597 4945 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/4f0c5f83cb1a3cc3daed2b69cf4d441db9f6e459b638907d31584174504a9a04/diff" to get inode usage: stat /var/lib/containers/storage/overlay/4f0c5f83cb1a3cc3daed2b69cf4d441db9f6e459b638907d31584174504a9a04/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_neutron-7b595c49d9-jngxp_85ab2778-4811-4700-bca0-d55c9282699f/neutron-api/0.log" to get inode usage: stat /var/log/pods/openstack_neutron-7b595c49d9-jngxp_85ab2778-4811-4700-bca0-d55c9282699f/neutron-api/0.log: no such file or directory Dec 06 07:21:14 crc kubenswrapper[4945]: I1206 07:21:14.286423 4945 scope.go:117] "RemoveContainer" containerID="c7eb1e87ecd225ee2282907be972c030861b9852e609a8c7badc9b0c36c11a22" Dec 06 07:21:14 crc kubenswrapper[4945]: I1206 07:21:14.312867 4945 scope.go:117] "RemoveContainer" containerID="bb3773d5b61628e5effd422bfa3be33b5be1e9ba5f96b66980463e518f4da6ac" Dec 06 07:21:14 crc kubenswrapper[4945]: I1206 07:21:14.343296 4945 scope.go:117] "RemoveContainer" containerID="ac89c0e1b702d43f1a10ceb5d1fa331abe2da4bec10b71837cf55f775bb4f9c7" Dec 06 07:21:16 crc kubenswrapper[4945]: E1206 07:21:16.133350 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:16 crc kubenswrapper[4945]: E1206 07:21:16.134907 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:16 crc kubenswrapper[4945]: E1206 07:21:16.135003 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:16 crc kubenswrapper[4945]: E1206 07:21:16.135715 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 06 07:21:16 crc kubenswrapper[4945]: E1206 07:21:16.135758 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" Dec 06 07:21:16 crc kubenswrapper[4945]: E1206 07:21:16.137182 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:16 crc kubenswrapper[4945]: E1206 07:21:16.139499 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 06 07:21:16 crc kubenswrapper[4945]: E1206 07:21:16.139597 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2996k" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovs-vswitchd" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.052734 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2996k_9c092be9-37db-49bb-be6b-56ce2e0e15fd/ovs-vswitchd/0.log" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.054423 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.164986 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bl4s\" (UniqueName: \"kubernetes.io/projected/9c092be9-37db-49bb-be6b-56ce2e0e15fd-kube-api-access-5bl4s\") pod \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.165106 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c092be9-37db-49bb-be6b-56ce2e0e15fd-scripts\") pod \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.165200 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-run\") pod \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.165307 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-lib\") pod \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.165380 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-etc-ovs\") pod \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.165396 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-run" (OuterVolumeSpecName: "var-run") pod "9c092be9-37db-49bb-be6b-56ce2e0e15fd" (UID: "9c092be9-37db-49bb-be6b-56ce2e0e15fd"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.165436 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-log\") pod \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\" (UID: \"9c092be9-37db-49bb-be6b-56ce2e0e15fd\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.165484 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "9c092be9-37db-49bb-be6b-56ce2e0e15fd" (UID: "9c092be9-37db-49bb-be6b-56ce2e0e15fd"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.165491 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-log" (OuterVolumeSpecName: "var-log") pod "9c092be9-37db-49bb-be6b-56ce2e0e15fd" (UID: "9c092be9-37db-49bb-be6b-56ce2e0e15fd"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.165434 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-lib" (OuterVolumeSpecName: "var-lib") pod "9c092be9-37db-49bb-be6b-56ce2e0e15fd" (UID: "9c092be9-37db-49bb-be6b-56ce2e0e15fd"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.166266 4945 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-run\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.166313 4945 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-lib\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.166326 4945 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-etc-ovs\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.166340 4945 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9c092be9-37db-49bb-be6b-56ce2e0e15fd-var-log\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.166580 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c092be9-37db-49bb-be6b-56ce2e0e15fd-scripts" (OuterVolumeSpecName: "scripts") pod "9c092be9-37db-49bb-be6b-56ce2e0e15fd" (UID: "9c092be9-37db-49bb-be6b-56ce2e0e15fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.171197 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c092be9-37db-49bb-be6b-56ce2e0e15fd-kube-api-access-5bl4s" (OuterVolumeSpecName: "kube-api-access-5bl4s") pod "9c092be9-37db-49bb-be6b-56ce2e0e15fd" (UID: "9c092be9-37db-49bb-be6b-56ce2e0e15fd"). InnerVolumeSpecName "kube-api-access-5bl4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.278706 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bl4s\" (UniqueName: \"kubernetes.io/projected/9c092be9-37db-49bb-be6b-56ce2e0e15fd-kube-api-access-5bl4s\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.278774 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c092be9-37db-49bb-be6b-56ce2e0e15fd-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.487457 4945 generic.go:334] "Generic (PLEG): container finished" podID="af58797c-3af9-4731-88dc-5c03e52217b4" containerID="789049d9004e9570b8c0301ec30ded93adc7a99fef755791d3e1828734190e19" exitCode=137 Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.487532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"789049d9004e9570b8c0301ec30ded93adc7a99fef755791d3e1828734190e19"} Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.491739 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2996k_9c092be9-37db-49bb-be6b-56ce2e0e15fd/ovs-vswitchd/0.log" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.493058 4945 generic.go:334] "Generic (PLEG): container finished" podID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" exitCode=137 Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.493099 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2996k" event={"ID":"9c092be9-37db-49bb-be6b-56ce2e0e15fd","Type":"ContainerDied","Data":"59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600"} Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.493135 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2996k" event={"ID":"9c092be9-37db-49bb-be6b-56ce2e0e15fd","Type":"ContainerDied","Data":"f5715ee34ed4cb57a22cbe9aa3687f0bce4b9d859a4437c104d0b5751af78ab7"} Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.493155 4945 scope.go:117] "RemoveContainer" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.493232 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2996k" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.517760 4945 scope.go:117] "RemoveContainer" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.530155 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-2996k"] Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.544836 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-2996k"] Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.557581 4945 scope.go:117] "RemoveContainer" containerID="febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.586774 4945 scope.go:117] "RemoveContainer" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" Dec 06 07:21:17 crc kubenswrapper[4945]: E1206 07:21:17.587169 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600\": container with ID starting with 59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600 not found: ID does not exist" containerID="59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.587268 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600"} err="failed to get container status \"59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600\": rpc error: code = NotFound desc = could not find container \"59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600\": container with ID starting with 59d07839f1dc0f8ae058d82ab0b65657a58e536f2974525850610836c3a56600 not found: ID does not exist" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.587338 4945 scope.go:117] "RemoveContainer" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" Dec 06 07:21:17 crc kubenswrapper[4945]: E1206 07:21:17.587571 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e\": container with ID starting with ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e not found: ID does not exist" containerID="ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.587601 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e"} err="failed to get container status \"ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e\": rpc error: code = NotFound desc = could not find container \"ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e\": container with ID starting with ca3ee823593675acbe830ed6aa62fa833066c513898196b4581977201d9a2f2e not found: ID does not exist" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.587618 4945 scope.go:117] "RemoveContainer" containerID="febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92" Dec 06 07:21:17 crc kubenswrapper[4945]: E1206 07:21:17.587842 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92\": container with ID starting with febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92 not found: ID does not exist" containerID="febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.587868 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92"} err="failed to get container status \"febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92\": rpc error: code = NotFound desc = could not find container \"febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92\": container with ID starting with febd122e1669e0133fdf6837fc5824c2ecd1aa9012073f3ea06fe56941abca92 not found: ID does not exist" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.739871 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.887440 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"af58797c-3af9-4731-88dc-5c03e52217b4\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.887502 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-cache\") pod \"af58797c-3af9-4731-88dc-5c03e52217b4\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.887564 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v99kc\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-kube-api-access-v99kc\") pod \"af58797c-3af9-4731-88dc-5c03e52217b4\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.887655 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-lock\") pod \"af58797c-3af9-4731-88dc-5c03e52217b4\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.887701 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") pod \"af58797c-3af9-4731-88dc-5c03e52217b4\" (UID: \"af58797c-3af9-4731-88dc-5c03e52217b4\") " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.888217 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-lock" (OuterVolumeSpecName: "lock") pod "af58797c-3af9-4731-88dc-5c03e52217b4" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.888363 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-cache" (OuterVolumeSpecName: "cache") pod "af58797c-3af9-4731-88dc-5c03e52217b4" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.893267 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "af58797c-3af9-4731-88dc-5c03e52217b4" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.893385 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-kube-api-access-v99kc" (OuterVolumeSpecName: "kube-api-access-v99kc") pod "af58797c-3af9-4731-88dc-5c03e52217b4" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4"). InnerVolumeSpecName "kube-api-access-v99kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.893435 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "swift") pod "af58797c-3af9-4731-88dc-5c03e52217b4" (UID: "af58797c-3af9-4731-88dc-5c03e52217b4"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.990123 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v99kc\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-kube-api-access-v99kc\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.990513 4945 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-lock\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.990524 4945 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af58797c-3af9-4731-88dc-5c03e52217b4-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.990580 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 06 07:21:17 crc kubenswrapper[4945]: I1206 07:21:17.990591 4945 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/af58797c-3af9-4731-88dc-5c03e52217b4-cache\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.004557 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.091858 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.511845 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af58797c-3af9-4731-88dc-5c03e52217b4","Type":"ContainerDied","Data":"88e498f32185431bf282d55d7dab63531bc10a134d582784ca5237d4afa2b48f"} Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.511896 4945 scope.go:117] "RemoveContainer" containerID="789049d9004e9570b8c0301ec30ded93adc7a99fef755791d3e1828734190e19" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.512729 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.544647 4945 scope.go:117] "RemoveContainer" containerID="ff4e063c23aaff06ab010a43b39361561a092ed784e7fd983facb5b882fd3225" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.548602 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.555647 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.560388 4945 scope.go:117] "RemoveContainer" containerID="6088cfa6a692765f885d01e25babb5993fb4320f6715c412196029c872327c41" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.575722 4945 scope.go:117] "RemoveContainer" containerID="e0fbc46abb12d20db2057256ae54a1bd2c5b3df4bd87a0d9a1700ac12c2ce2c1" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.591291 4945 scope.go:117] "RemoveContainer" containerID="0c80efac78458edb45ee60a60f64b441b635ee4a49351c959c569def27d0e665" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.608152 4945 scope.go:117] "RemoveContainer" containerID="a8901783d5561a470b55820422085e5ac6d8ccd2649a4eec8701a774fe804faf" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.624892 4945 scope.go:117] "RemoveContainer" containerID="04e71c5a4df960ad72d6476b9f52dc9c05815786d261a2816740f83faab2d771" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.642352 4945 scope.go:117] "RemoveContainer" containerID="de0927bc9e6dd18e1a79f623e51e8901bc80dec257144ca686e887a312f70a86" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.660379 4945 scope.go:117] "RemoveContainer" containerID="1991abb13d702f2e5d57933ab6d10cf0d4c07b1134ee92235420a55b874e1f9e" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.678859 4945 scope.go:117] "RemoveContainer" containerID="cbfe05d94d7270d286fed8b4840eccd1701ac59c60fb2b3214890ceec204a5cd" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.700124 4945 scope.go:117] "RemoveContainer" containerID="195d75cef2ef89afdb1df93e86f8e6e38bae58c3938182ede8e3b5daa3fde195" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.727003 4945 scope.go:117] "RemoveContainer" containerID="a0fcd13fd05e6559ba8dfeed754e555d93dc6f54fcefdb814974d89ffc3bac43" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.750385 4945 scope.go:117] "RemoveContainer" containerID="ff9b25f9862ebd3ee6236d1ae7fa95a0a44adbb1a4d9be0d4f84005b7298f8fd" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.772071 4945 scope.go:117] "RemoveContainer" containerID="57a847dca15d392b38d22b41bbbc40b779272394f028698e226e315d75751d47" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.793375 4945 scope.go:117] "RemoveContainer" containerID="8f302203e6126bf00e33188ea6dd1f7e034a8cbfe7c5182693c187c7f24b819a" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.970415 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" path="/var/lib/kubelet/pods/9c092be9-37db-49bb-be6b-56ce2e0e15fd/volumes" Dec 06 07:21:18 crc kubenswrapper[4945]: I1206 07:21:18.971234 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" path="/var/lib/kubelet/pods/af58797c-3af9-4731-88dc-5c03e52217b4/volumes" Dec 06 07:21:19 crc kubenswrapper[4945]: I1206 07:21:19.378501 4945 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podd5496b71-8389-4d16-97c0-5752a2f1ef2a"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podd5496b71-8389-4d16-97c0-5752a2f1ef2a] : Timed out while waiting for systemd to remove kubepods-besteffort-podd5496b71_8389_4d16_97c0_5752a2f1ef2a.slice" Dec 06 07:21:21 crc kubenswrapper[4945]: I1206 07:21:21.953030 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:21:21 crc kubenswrapper[4945]: E1206 07:21:21.953554 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:21:24 crc kubenswrapper[4945]: E1206 07:21:24.194579 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83eee8a2_f458_4193_8023_ca1228dc92a3.slice/crio-a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83eee8a2_f458_4193_8023_ca1228dc92a3.slice/crio-conmon-a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298.scope\": RecentStats: unable to find data in memory cache]" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.298166 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0216c-account-delete-bx5zs" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.306524 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapibadc-account-delete-4n95m" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.386151 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6fbf826-c8e6-4231-b479-90fe7de5a175-operator-scripts\") pod \"a6fbf826-c8e6-4231-b479-90fe7de5a175\" (UID: \"a6fbf826-c8e6-4231-b479-90fe7de5a175\") " Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.386412 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p6j5\" (UniqueName: \"kubernetes.io/projected/a6fbf826-c8e6-4231-b479-90fe7de5a175-kube-api-access-4p6j5\") pod \"a6fbf826-c8e6-4231-b479-90fe7de5a175\" (UID: \"a6fbf826-c8e6-4231-b479-90fe7de5a175\") " Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.387051 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6fbf826-c8e6-4231-b479-90fe7de5a175-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6fbf826-c8e6-4231-b479-90fe7de5a175" (UID: "a6fbf826-c8e6-4231-b479-90fe7de5a175"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.391568 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6fbf826-c8e6-4231-b479-90fe7de5a175-kube-api-access-4p6j5" (OuterVolumeSpecName: "kube-api-access-4p6j5") pod "a6fbf826-c8e6-4231-b479-90fe7de5a175" (UID: "a6fbf826-c8e6-4231-b479-90fe7de5a175"). InnerVolumeSpecName "kube-api-access-4p6j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.488229 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggwb9\" (UniqueName: \"kubernetes.io/projected/83eee8a2-f458-4193-8023-ca1228dc92a3-kube-api-access-ggwb9\") pod \"83eee8a2-f458-4193-8023-ca1228dc92a3\" (UID: \"83eee8a2-f458-4193-8023-ca1228dc92a3\") " Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.488386 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83eee8a2-f458-4193-8023-ca1228dc92a3-operator-scripts\") pod \"83eee8a2-f458-4193-8023-ca1228dc92a3\" (UID: \"83eee8a2-f458-4193-8023-ca1228dc92a3\") " Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.488767 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p6j5\" (UniqueName: \"kubernetes.io/projected/a6fbf826-c8e6-4231-b479-90fe7de5a175-kube-api-access-4p6j5\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.488793 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6fbf826-c8e6-4231-b479-90fe7de5a175-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.488930 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83eee8a2-f458-4193-8023-ca1228dc92a3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83eee8a2-f458-4193-8023-ca1228dc92a3" (UID: "83eee8a2-f458-4193-8023-ca1228dc92a3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.493914 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83eee8a2-f458-4193-8023-ca1228dc92a3-kube-api-access-ggwb9" (OuterVolumeSpecName: "kube-api-access-ggwb9") pod "83eee8a2-f458-4193-8023-ca1228dc92a3" (UID: "83eee8a2-f458-4193-8023-ca1228dc92a3"). InnerVolumeSpecName "kube-api-access-ggwb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.589679 4945 generic.go:334] "Generic (PLEG): container finished" podID="a6fbf826-c8e6-4231-b479-90fe7de5a175" containerID="37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da" exitCode=137 Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.589767 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0216c-account-delete-bx5zs" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.589760 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0216c-account-delete-bx5zs" event={"ID":"a6fbf826-c8e6-4231-b479-90fe7de5a175","Type":"ContainerDied","Data":"37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da"} Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.589852 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0216c-account-delete-bx5zs" event={"ID":"a6fbf826-c8e6-4231-b479-90fe7de5a175","Type":"ContainerDied","Data":"2aceb45c5bf7d83622fb1a78f07382add9a1018ebff2cef376918ef0889ca602"} Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.589924 4945 scope.go:117] "RemoveContainer" containerID="37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.592022 4945 generic.go:334] "Generic (PLEG): container finished" podID="83eee8a2-f458-4193-8023-ca1228dc92a3" containerID="a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298" exitCode=137 Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.592054 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapibadc-account-delete-4n95m" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.592064 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapibadc-account-delete-4n95m" event={"ID":"83eee8a2-f458-4193-8023-ca1228dc92a3","Type":"ContainerDied","Data":"a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298"} Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.592095 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapibadc-account-delete-4n95m" event={"ID":"83eee8a2-f458-4193-8023-ca1228dc92a3","Type":"ContainerDied","Data":"e1b7abd787f3f862aa8ae56e6c3ece46e8021c2d3eb8f17e25be9eec176571dd"} Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.593296 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83eee8a2-f458-4193-8023-ca1228dc92a3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.593457 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggwb9\" (UniqueName: \"kubernetes.io/projected/83eee8a2-f458-4193-8023-ca1228dc92a3-kube-api-access-ggwb9\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.627031 4945 scope.go:117] "RemoveContainer" containerID="37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da" Dec 06 07:21:24 crc kubenswrapper[4945]: E1206 07:21:24.627618 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da\": container with ID starting with 37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da not found: ID does not exist" containerID="37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.627665 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da"} err="failed to get container status \"37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da\": rpc error: code = NotFound desc = could not find container \"37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da\": container with ID starting with 37d3991d40bf6f1076687b5f525ccf69350a3d8e13c5ad451a76140fa71990da not found: ID does not exist" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.627689 4945 scope.go:117] "RemoveContainer" containerID="a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.644861 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapibadc-account-delete-4n95m"] Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.653744 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapibadc-account-delete-4n95m"] Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.663679 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0216c-account-delete-bx5zs"] Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.671086 4945 scope.go:117] "RemoveContainer" containerID="a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298" Dec 06 07:21:24 crc kubenswrapper[4945]: E1206 07:21:24.671555 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298\": container with ID starting with a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298 not found: ID does not exist" containerID="a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.671601 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298"} err="failed to get container status \"a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298\": rpc error: code = NotFound desc = could not find container \"a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298\": container with ID starting with a28779fdba9dc8d82745c454b82aa89fd2f01ccb231fb0db3958a507eba1b298 not found: ID does not exist" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.672976 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0216c-account-delete-bx5zs"] Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.963073 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83eee8a2-f458-4193-8023-ca1228dc92a3" path="/var/lib/kubelet/pods/83eee8a2-f458-4193-8023-ca1228dc92a3/volumes" Dec 06 07:21:24 crc kubenswrapper[4945]: I1206 07:21:24.963811 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6fbf826-c8e6-4231-b479-90fe7de5a175" path="/var/lib/kubelet/pods/a6fbf826-c8e6-4231-b479-90fe7de5a175/volumes" Dec 06 07:21:25 crc kubenswrapper[4945]: E1206 07:21:25.814782 4945 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 06 07:21:25 crc kubenswrapper[4945]: E1206 07:21:25.814897 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts podName:4231a0fc-b0c8-48da-8fab-75cd817dd3ef nodeName:}" failed. No retries permitted until 2025-12-06 07:21:57.814863913 +0000 UTC m=+1731.269724977 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts") pod "neutron975b-account-delete-nvsg5" (UID: "4231a0fc-b0c8-48da-8fab-75cd817dd3ef") : configmap "openstack-scripts" not found Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.616213 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron975b-account-delete-nvsg5" Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.623720 4945 generic.go:334] "Generic (PLEG): container finished" podID="4231a0fc-b0c8-48da-8fab-75cd817dd3ef" containerID="0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c" exitCode=137 Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.623767 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron975b-account-delete-nvsg5" event={"ID":"4231a0fc-b0c8-48da-8fab-75cd817dd3ef","Type":"ContainerDied","Data":"0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c"} Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.623992 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron975b-account-delete-nvsg5" event={"ID":"4231a0fc-b0c8-48da-8fab-75cd817dd3ef","Type":"ContainerDied","Data":"3656e7d4d7212ac866e8c54991cfcaf41873338da41de2cded7fb37a11f86870"} Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.624028 4945 scope.go:117] "RemoveContainer" containerID="0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c" Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.624029 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron975b-account-delete-nvsg5" Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.663119 4945 scope.go:117] "RemoveContainer" containerID="0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c" Dec 06 07:21:26 crc kubenswrapper[4945]: E1206 07:21:26.664440 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c\": container with ID starting with 0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c not found: ID does not exist" containerID="0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c" Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.664503 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c"} err="failed to get container status \"0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c\": rpc error: code = NotFound desc = could not find container \"0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c\": container with ID starting with 0b209b1dfffd677c848b395dfb848c633f9187f0a6d67e502f62d9872c47369c not found: ID does not exist" Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.726745 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfj5g\" (UniqueName: \"kubernetes.io/projected/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-kube-api-access-kfj5g\") pod \"4231a0fc-b0c8-48da-8fab-75cd817dd3ef\" (UID: \"4231a0fc-b0c8-48da-8fab-75cd817dd3ef\") " Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.726870 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts\") pod \"4231a0fc-b0c8-48da-8fab-75cd817dd3ef\" (UID: \"4231a0fc-b0c8-48da-8fab-75cd817dd3ef\") " Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.727826 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4231a0fc-b0c8-48da-8fab-75cd817dd3ef" (UID: "4231a0fc-b0c8-48da-8fab-75cd817dd3ef"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.731913 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-kube-api-access-kfj5g" (OuterVolumeSpecName: "kube-api-access-kfj5g") pod "4231a0fc-b0c8-48da-8fab-75cd817dd3ef" (UID: "4231a0fc-b0c8-48da-8fab-75cd817dd3ef"). InnerVolumeSpecName "kube-api-access-kfj5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.828394 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.828434 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfj5g\" (UniqueName: \"kubernetes.io/projected/4231a0fc-b0c8-48da-8fab-75cd817dd3ef-kube-api-access-kfj5g\") on node \"crc\" DevicePath \"\"" Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.967634 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron975b-account-delete-nvsg5"] Dec 06 07:21:26 crc kubenswrapper[4945]: I1206 07:21:26.970353 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron975b-account-delete-nvsg5"] Dec 06 07:21:28 crc kubenswrapper[4945]: I1206 07:21:28.961966 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4231a0fc-b0c8-48da-8fab-75cd817dd3ef" path="/var/lib/kubelet/pods/4231a0fc-b0c8-48da-8fab-75cd817dd3ef/volumes" Dec 06 07:21:32 crc kubenswrapper[4945]: I1206 07:21:32.953589 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:21:32 crc kubenswrapper[4945]: E1206 07:21:32.954097 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.141673 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.142560 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227014 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kn58l"] Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227580 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ab2778-4811-4700-bca0-d55c9282699f" containerName="neutron-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227598 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ab2778-4811-4700-bca0-d55c9282699f" containerName="neutron-api" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227607 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" containerName="galera" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227614 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" containerName="galera" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227628 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" containerName="glance-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227635 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" containerName="glance-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227646 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" containerName="rabbitmq" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227652 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" containerName="rabbitmq" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227658 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" containerName="barbican-keystone-listener-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227664 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" containerName="barbican-keystone-listener-log" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227672 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="ceilometer-notification-agent" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227688 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="ceilometer-notification-agent" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227701 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232baa8c-2d18-49f3-b230-362181e687df" containerName="openstack-network-exporter" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227707 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="232baa8c-2d18-49f3-b230-362181e687df" containerName="openstack-network-exporter" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227715 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-server" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227721 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-server" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227731 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-replicator" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227737 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-replicator" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227747 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="250c85e2-e14d-4f93-822c-c9a88352eafe" containerName="rabbitmq" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227759 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="250c85e2-e14d-4f93-822c-c9a88352eafe" containerName="rabbitmq" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227766 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72607de1-4f23-441e-b722-3844b2eb6d15" containerName="nova-cell0-conductor-conductor" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227772 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72607de1-4f23-441e-b722-3844b2eb6d15" containerName="nova-cell0-conductor-conductor" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227783 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5228457-326b-4470-ad26-f7bba1f50e18" containerName="kube-state-metrics" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227788 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5228457-326b-4470-ad26-f7bba1f50e18" containerName="kube-state-metrics" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227799 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe30cfa7-9746-4771-bac5-29e3a3662aec" containerName="galera" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227804 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe30cfa7-9746-4771-bac5-29e3a3662aec" containerName="galera" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227814 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-expirer" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227820 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-expirer" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227832 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1cb89b2-7918-457d-a36a-2cec380f3854" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227837 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1cb89b2-7918-457d-a36a-2cec380f3854" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227847 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="swift-recon-cron" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227853 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="swift-recon-cron" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227860 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="ceilometer-central-agent" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227866 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="ceilometer-central-agent" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227875 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe30cfa7-9746-4771-bac5-29e3a3662aec" containerName="mysql-bootstrap" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227880 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe30cfa7-9746-4771-bac5-29e3a3662aec" containerName="mysql-bootstrap" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227888 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83eee8a2-f458-4193-8023-ca1228dc92a3" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227893 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="83eee8a2-f458-4193-8023-ca1228dc92a3" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227902 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" containerName="ovn-controller" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227909 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" containerName="ovn-controller" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227916 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6fbf826-c8e6-4231-b479-90fe7de5a175" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227922 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6fbf826-c8e6-4231-b479-90fe7de5a175" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227932 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="569f0e1c-e58f-4e87-bde3-e7be4659639a" containerName="placement-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227937 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="569f0e1c-e58f-4e87-bde3-e7be4659639a" containerName="placement-log" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227945 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" containerName="mysql-bootstrap" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227951 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" containerName="mysql-bootstrap" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227962 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server-init" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227968 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server-init" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227976 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovs-vswitchd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227981 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovs-vswitchd" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.227991 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-reaper" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.227996 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-reaper" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228006 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-replicator" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228011 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-replicator" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228019 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerName="cinder-api-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228025 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerName="cinder-api-log" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228031 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-replicator" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228037 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-replicator" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228043 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="rsync" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228048 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="rsync" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228058 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" containerName="glance-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228063 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" containerName="glance-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228071 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-server" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228077 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-server" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228088 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" containerName="barbican-worker-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228094 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" containerName="barbican-worker-log" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228104 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22eb2234-b14b-474c-b1bf-cf726acb064b" containerName="nova-cell1-conductor-conductor" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228110 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="22eb2234-b14b-474c-b1bf-cf726acb064b" containerName="nova-cell1-conductor-conductor" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228123 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" containerName="cinder-scheduler" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228128 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" containerName="cinder-scheduler" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228136 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" containerName="setup-container" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228141 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" containerName="setup-container" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228149 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" containerName="probe" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228154 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" containerName="probe" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228163 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e3c48ac-c191-4478-877e-9600b6f63929" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228168 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e3c48ac-c191-4478-877e-9600b6f63929" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228179 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerName="nova-api-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228185 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerName="nova-api-log" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228194 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-auditor" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228199 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-auditor" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228207 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-auditor" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228212 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-auditor" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228219 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="proxy-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228224 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="proxy-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228232 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerName="barbican-api-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228238 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerName="barbican-api-log" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228247 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-metadata" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228253 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-metadata" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228262 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4231a0fc-b0c8-48da-8fab-75cd817dd3ef" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228268 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4231a0fc-b0c8-48da-8fab-75cd817dd3ef" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228289 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" containerName="barbican-worker" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228295 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" containerName="barbican-worker" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228306 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="250c85e2-e14d-4f93-822c-c9a88352eafe" containerName="setup-container" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228312 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="250c85e2-e14d-4f93-822c-c9a88352eafe" containerName="setup-container" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228321 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerName="nova-api-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228327 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerName="nova-api-api" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228338 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95dc303d-9736-4c29-a1f4-f8b362179eeb" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228344 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="95dc303d-9736-4c29-a1f4-f8b362179eeb" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228352 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerName="barbican-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228358 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerName="barbican-api" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228368 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-server" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228373 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-server" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228382 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228388 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228397 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" containerName="memcached" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228403 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" containerName="memcached" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228411 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerName="cinder-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228416 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerName="cinder-api" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228423 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-updater" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228428 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-updater" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228435 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="569f0e1c-e58f-4e87-bde3-e7be4659639a" containerName="placement-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228441 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="569f0e1c-e58f-4e87-bde3-e7be4659639a" containerName="placement-api" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228451 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228456 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-log" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228466 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="sg-core" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228474 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="sg-core" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228481 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232baa8c-2d18-49f3-b230-362181e687df" containerName="ovn-northd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228488 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="232baa8c-2d18-49f3-b230-362181e687df" containerName="ovn-northd" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228500 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-updater" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228508 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-updater" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228518 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae75a6a-2656-4e65-a84d-8767a570ce8f" containerName="nova-scheduler-scheduler" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228525 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae75a6a-2656-4e65-a84d-8767a570ce8f" containerName="nova-scheduler-scheduler" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228537 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228543 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228553 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" containerName="glance-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228561 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" containerName="glance-log" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228572 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-auditor" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228577 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-auditor" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228586 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ab2778-4811-4700-bca0-d55c9282699f" containerName="neutron-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228591 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ab2778-4811-4700-bca0-d55c9282699f" containerName="neutron-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228598 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" containerName="barbican-keystone-listener" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228604 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" containerName="barbican-keystone-listener" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228611 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af85716-fd7e-45e9-8320-4db8800302a8" containerName="keystone-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228616 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af85716-fd7e-45e9-8320-4db8800302a8" containerName="keystone-api" Dec 06 07:21:49 crc kubenswrapper[4945]: E1206 07:21:49.228625 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" containerName="glance-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228630 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" containerName="glance-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228759 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d71d0084-b4bc-4be5-971b-eced6e6d66ea" containerName="rabbitmq" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228771 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" containerName="cinder-scheduler" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228779 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af85716-fd7e-45e9-8320-4db8800302a8" containerName="keystone-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228784 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fb32cca-c8f2-4ff6-a05c-25f3e56b5478" containerName="ovn-controller" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228794 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" containerName="barbican-worker-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228802 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="sg-core" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228808 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-server" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228817 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ab2778-4811-4700-bca0-d55c9282699f" containerName="neutron-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228826 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" containerName="glance-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228835 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerName="barbican-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228844 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="232baa8c-2d18-49f3-b230-362181e687df" containerName="openstack-network-exporter" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228853 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="232baa8c-2d18-49f3-b230-362181e687df" containerName="ovn-northd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228861 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="250c85e2-e14d-4f93-822c-c9a88352eafe" containerName="rabbitmq" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228870 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="569f0e1c-e58f-4e87-bde3-e7be4659639a" containerName="placement-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228877 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="proxy-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228885 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf7f84bc-3d5c-40b1-ad81-c52f1cb8271b" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228895 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-replicator" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228904 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" containerName="barbican-keystone-listener-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228915 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-replicator" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228924 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5c5a14-2ccf-46b4-adbb-3ad2ab0a6975" containerName="barbican-api-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228933 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72607de1-4f23-441e-b722-3844b2eb6d15" containerName="nova-cell0-conductor-conductor" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228940 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="ceilometer-central-agent" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228947 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef14eff5-a3d8-48a8-b3fc-b59b97f68b67" containerName="memcached" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228953 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8dae45-f01e-4ef0-8814-07511ff2fcad" containerName="galera" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228961 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerName="cinder-api-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228967 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovsdb-server" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228974 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="569f0e1c-e58f-4e87-bde3-e7be4659639a" containerName="placement-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228980 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8fbb8a7-4226-4b2f-9e95-621059e6283d" containerName="ceilometer-notification-agent" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228986 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-expirer" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.228994 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6fbf826-c8e6-4231-b479-90fe7de5a175" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229002 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-replicator" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229009 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-server" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229017 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="edbffd82-c356-43a3-94a3-3549b7d9cd31" containerName="cinder-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229023 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" containerName="glance-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229032 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ba3e057-d321-48ca-b7a0-8d791cbb7169" containerName="glance-httpd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229041 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-auditor" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229049 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d2b3b4-2d39-4966-b16c-00e3b9fdbeac" containerName="barbican-worker" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229057 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="966c9b21-65c4-4d2a-af0f-0041a9bfea08" containerName="barbican-keystone-listener" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229066 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="rsync" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229073 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-metadata" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229079 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="95dc303d-9736-4c29-a1f4-f8b362179eeb" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229090 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e3c48ac-c191-4478-877e-9600b6f63929" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229102 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4231a0fc-b0c8-48da-8fab-75cd817dd3ef" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229114 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="22eb2234-b14b-474c-b1bf-cf726acb064b" containerName="nova-cell1-conductor-conductor" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229122 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae75a6a-2656-4e65-a84d-8767a570ce8f" containerName="nova-scheduler-scheduler" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229131 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="swift-recon-cron" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229138 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-updater" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229144 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-auditor" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229152 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="object-updater" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229160 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5228457-326b-4470-ad26-f7bba1f50e18" containerName="kube-state-metrics" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229167 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe30cfa7-9746-4771-bac5-29e3a3662aec" containerName="galera" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229174 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerName="nova-api-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229182 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9296c3b2-7997-45a0-a7e9-bc1f45ec204e" containerName="nova-api-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229189 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-auditor" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229198 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ab2778-4811-4700-bca0-d55c9282699f" containerName="neutron-api" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229206 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="efad5b26-d064-4d06-8cc8-361d094dfa9c" containerName="nova-metadata-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229215 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c092be9-37db-49bb-be6b-56ce2e0e15fd" containerName="ovs-vswitchd" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229223 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1cb89b2-7918-457d-a36a-2cec380f3854" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229233 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="account-reaper" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229240 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d74c1f-98d7-40ac-a1e8-233604971bc9" containerName="probe" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229247 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf0fa4a-4a9c-4644-9fe3-413500c59a98" containerName="glance-log" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229254 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58797c-3af9-4731-88dc-5c03e52217b4" containerName="container-server" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.229263 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="83eee8a2-f458-4193-8023-ca1228dc92a3" containerName="mariadb-account-delete" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.230371 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.262876 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kn58l"] Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.313016 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-utilities\") pod \"redhat-operators-kn58l\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.313161 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-catalog-content\") pod \"redhat-operators-kn58l\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.313227 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56vms\" (UniqueName: \"kubernetes.io/projected/a1ca9fe2-4986-4e85-838b-172332c58972-kube-api-access-56vms\") pod \"redhat-operators-kn58l\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.414663 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-catalog-content\") pod \"redhat-operators-kn58l\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.414716 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56vms\" (UniqueName: \"kubernetes.io/projected/a1ca9fe2-4986-4e85-838b-172332c58972-kube-api-access-56vms\") pod \"redhat-operators-kn58l\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.414788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-utilities\") pod \"redhat-operators-kn58l\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.415247 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-catalog-content\") pod \"redhat-operators-kn58l\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.415359 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-utilities\") pod \"redhat-operators-kn58l\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.435747 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56vms\" (UniqueName: \"kubernetes.io/projected/a1ca9fe2-4986-4e85-838b-172332c58972-kube-api-access-56vms\") pod \"redhat-operators-kn58l\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:49 crc kubenswrapper[4945]: I1206 07:21:49.575231 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:50 crc kubenswrapper[4945]: I1206 07:21:50.009774 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kn58l"] Dec 06 07:21:50 crc kubenswrapper[4945]: I1206 07:21:50.269652 4945 generic.go:334] "Generic (PLEG): container finished" podID="a1ca9fe2-4986-4e85-838b-172332c58972" containerID="800c3cb600e81ee8207b7eccd843d992111a19b03538ae88cdd9d046fa65586a" exitCode=0 Dec 06 07:21:50 crc kubenswrapper[4945]: I1206 07:21:50.269700 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn58l" event={"ID":"a1ca9fe2-4986-4e85-838b-172332c58972","Type":"ContainerDied","Data":"800c3cb600e81ee8207b7eccd843d992111a19b03538ae88cdd9d046fa65586a"} Dec 06 07:21:50 crc kubenswrapper[4945]: I1206 07:21:50.269951 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn58l" event={"ID":"a1ca9fe2-4986-4e85-838b-172332c58972","Type":"ContainerStarted","Data":"6759242cd9023f2832b225c552dbc1be70ea67f73e6d0a33eb48a7b55a7a097e"} Dec 06 07:21:52 crc kubenswrapper[4945]: I1206 07:21:52.294737 4945 generic.go:334] "Generic (PLEG): container finished" podID="a1ca9fe2-4986-4e85-838b-172332c58972" containerID="7ab0661c1687f85b82ca3f68f6024b6a8cc7d4f374e818cfc6af141dd3e204b8" exitCode=0 Dec 06 07:21:52 crc kubenswrapper[4945]: I1206 07:21:52.294825 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn58l" event={"ID":"a1ca9fe2-4986-4e85-838b-172332c58972","Type":"ContainerDied","Data":"7ab0661c1687f85b82ca3f68f6024b6a8cc7d4f374e818cfc6af141dd3e204b8"} Dec 06 07:21:53 crc kubenswrapper[4945]: I1206 07:21:53.306543 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn58l" event={"ID":"a1ca9fe2-4986-4e85-838b-172332c58972","Type":"ContainerStarted","Data":"70d45fd603791060948f32156bbebf5e62bcbc8d41da5d6dd1d273b4b64986ad"} Dec 06 07:21:53 crc kubenswrapper[4945]: I1206 07:21:53.333314 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kn58l" podStartSLOduration=1.907031003 podStartE2EDuration="4.333296899s" podCreationTimestamp="2025-12-06 07:21:49 +0000 UTC" firstStartedPulling="2025-12-06 07:21:50.270944691 +0000 UTC m=+1723.725805735" lastFinishedPulling="2025-12-06 07:21:52.697210587 +0000 UTC m=+1726.152071631" observedRunningTime="2025-12-06 07:21:53.326481855 +0000 UTC m=+1726.781342939" watchObservedRunningTime="2025-12-06 07:21:53.333296899 +0000 UTC m=+1726.788157953" Dec 06 07:21:59 crc kubenswrapper[4945]: I1206 07:21:59.576001 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:59 crc kubenswrapper[4945]: I1206 07:21:59.576421 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:21:59 crc kubenswrapper[4945]: I1206 07:21:59.620698 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:22:00 crc kubenswrapper[4945]: I1206 07:22:00.430749 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:22:00 crc kubenswrapper[4945]: I1206 07:22:00.476790 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kn58l"] Dec 06 07:22:02 crc kubenswrapper[4945]: I1206 07:22:02.403299 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kn58l" podUID="a1ca9fe2-4986-4e85-838b-172332c58972" containerName="registry-server" containerID="cri-o://70d45fd603791060948f32156bbebf5e62bcbc8d41da5d6dd1d273b4b64986ad" gracePeriod=2 Dec 06 07:22:03 crc kubenswrapper[4945]: I1206 07:22:03.954362 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:22:03 crc kubenswrapper[4945]: E1206 07:22:03.954821 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.423797 4945 generic.go:334] "Generic (PLEG): container finished" podID="a1ca9fe2-4986-4e85-838b-172332c58972" containerID="70d45fd603791060948f32156bbebf5e62bcbc8d41da5d6dd1d273b4b64986ad" exitCode=0 Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.423839 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn58l" event={"ID":"a1ca9fe2-4986-4e85-838b-172332c58972","Type":"ContainerDied","Data":"70d45fd603791060948f32156bbebf5e62bcbc8d41da5d6dd1d273b4b64986ad"} Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.634378 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.728148 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56vms\" (UniqueName: \"kubernetes.io/projected/a1ca9fe2-4986-4e85-838b-172332c58972-kube-api-access-56vms\") pod \"a1ca9fe2-4986-4e85-838b-172332c58972\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.728481 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-utilities\") pod \"a1ca9fe2-4986-4e85-838b-172332c58972\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.728541 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-catalog-content\") pod \"a1ca9fe2-4986-4e85-838b-172332c58972\" (UID: \"a1ca9fe2-4986-4e85-838b-172332c58972\") " Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.729622 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-utilities" (OuterVolumeSpecName: "utilities") pod "a1ca9fe2-4986-4e85-838b-172332c58972" (UID: "a1ca9fe2-4986-4e85-838b-172332c58972"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.736581 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ca9fe2-4986-4e85-838b-172332c58972-kube-api-access-56vms" (OuterVolumeSpecName: "kube-api-access-56vms") pod "a1ca9fe2-4986-4e85-838b-172332c58972" (UID: "a1ca9fe2-4986-4e85-838b-172332c58972"). InnerVolumeSpecName "kube-api-access-56vms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.829568 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56vms\" (UniqueName: \"kubernetes.io/projected/a1ca9fe2-4986-4e85-838b-172332c58972-kube-api-access-56vms\") on node \"crc\" DevicePath \"\"" Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.829605 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.866696 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1ca9fe2-4986-4e85-838b-172332c58972" (UID: "a1ca9fe2-4986-4e85-838b-172332c58972"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:22:04 crc kubenswrapper[4945]: I1206 07:22:04.931346 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1ca9fe2-4986-4e85-838b-172332c58972-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:22:05 crc kubenswrapper[4945]: I1206 07:22:05.433920 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn58l" event={"ID":"a1ca9fe2-4986-4e85-838b-172332c58972","Type":"ContainerDied","Data":"6759242cd9023f2832b225c552dbc1be70ea67f73e6d0a33eb48a7b55a7a097e"} Dec 06 07:22:05 crc kubenswrapper[4945]: I1206 07:22:05.433981 4945 scope.go:117] "RemoveContainer" containerID="70d45fd603791060948f32156bbebf5e62bcbc8d41da5d6dd1d273b4b64986ad" Dec 06 07:22:05 crc kubenswrapper[4945]: I1206 07:22:05.434053 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kn58l" Dec 06 07:22:05 crc kubenswrapper[4945]: I1206 07:22:05.460316 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kn58l"] Dec 06 07:22:05 crc kubenswrapper[4945]: I1206 07:22:05.466058 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kn58l"] Dec 06 07:22:05 crc kubenswrapper[4945]: I1206 07:22:05.469704 4945 scope.go:117] "RemoveContainer" containerID="7ab0661c1687f85b82ca3f68f6024b6a8cc7d4f374e818cfc6af141dd3e204b8" Dec 06 07:22:05 crc kubenswrapper[4945]: I1206 07:22:05.497565 4945 scope.go:117] "RemoveContainer" containerID="800c3cb600e81ee8207b7eccd843d992111a19b03538ae88cdd9d046fa65586a" Dec 06 07:22:06 crc kubenswrapper[4945]: I1206 07:22:06.964212 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1ca9fe2-4986-4e85-838b-172332c58972" path="/var/lib/kubelet/pods/a1ca9fe2-4986-4e85-838b-172332c58972/volumes" Dec 06 07:22:14 crc kubenswrapper[4945]: I1206 07:22:14.696828 4945 scope.go:117] "RemoveContainer" containerID="29cdfb50edd3a43da06ed0091aa1d96671bf431d904eb9c175dff12ba56e0477" Dec 06 07:22:14 crc kubenswrapper[4945]: I1206 07:22:14.720225 4945 scope.go:117] "RemoveContainer" containerID="27630bdcc77303b93d309a9fbb7dc77578d594516e2a965d251b84f4c0aa821f" Dec 06 07:22:14 crc kubenswrapper[4945]: I1206 07:22:14.784675 4945 scope.go:117] "RemoveContainer" containerID="6af3e0e431d8760ff4b89dcefc9f22ce56f5f677ccbf93f887d5053936a15a72" Dec 06 07:22:14 crc kubenswrapper[4945]: I1206 07:22:14.809254 4945 scope.go:117] "RemoveContainer" containerID="e01ace9d4e5e9d743ae9af7802df7c36ec09040535d59f26854b256f8a64cc9b" Dec 06 07:22:14 crc kubenswrapper[4945]: I1206 07:22:14.845960 4945 scope.go:117] "RemoveContainer" containerID="3624afb9c68cb040abd682f4ced6ee22803012bbc4d7d8b26dd3e7ce553edb4c" Dec 06 07:22:14 crc kubenswrapper[4945]: I1206 07:22:14.871125 4945 scope.go:117] "RemoveContainer" containerID="b300df8609ad712283eee0541bfa67913ab230d553149ead1b538a7e5cf6a1d0" Dec 06 07:22:14 crc kubenswrapper[4945]: I1206 07:22:14.895726 4945 scope.go:117] "RemoveContainer" containerID="64091da47e5612c87b8350905f3c5270e29715775372b5e9af40732fcacee0da" Dec 06 07:22:14 crc kubenswrapper[4945]: I1206 07:22:14.936119 4945 scope.go:117] "RemoveContainer" containerID="1d208203ca7f66b7e885206b007a585ace9805113b2f6c3f8842bd6232730d75" Dec 06 07:22:14 crc kubenswrapper[4945]: I1206 07:22:14.977878 4945 scope.go:117] "RemoveContainer" containerID="1bfccd71aa9901558b2bdd7c5eb7435e28b26696e7097a7b152a71a444c574b1" Dec 06 07:22:14 crc kubenswrapper[4945]: I1206 07:22:14.997073 4945 scope.go:117] "RemoveContainer" containerID="328607c02897e7568c645c5ee8ef89528ebfeb9ca65456bea19a7c18881606e6" Dec 06 07:22:15 crc kubenswrapper[4945]: I1206 07:22:15.017976 4945 scope.go:117] "RemoveContainer" containerID="0503309da1afde1f127ec513bd61decd36e94ee9c302693c769b55d708825aeb" Dec 06 07:22:15 crc kubenswrapper[4945]: I1206 07:22:15.038008 4945 scope.go:117] "RemoveContainer" containerID="948373ff86352f7383fc745821f8ce1cfcddc2703574fde9f0f77d1c9931317f" Dec 06 07:22:15 crc kubenswrapper[4945]: I1206 07:22:15.075097 4945 scope.go:117] "RemoveContainer" containerID="40ab40310e6772f4b9f2541373cf860b3422560fe3fcc29e1733da457cd5d585" Dec 06 07:22:15 crc kubenswrapper[4945]: I1206 07:22:15.100828 4945 scope.go:117] "RemoveContainer" containerID="453ba68949233d03da1550a4fe35605ece36d1e51233e8fd776fbc8f82477d6c" Dec 06 07:22:15 crc kubenswrapper[4945]: I1206 07:22:15.120923 4945 scope.go:117] "RemoveContainer" containerID="3bd788957d65014f6df7ecdbf42af33341027a132b2359cc1089be7edcca1dc3" Dec 06 07:22:16 crc kubenswrapper[4945]: I1206 07:22:16.958024 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:22:16 crc kubenswrapper[4945]: E1206 07:22:16.959167 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:22:30 crc kubenswrapper[4945]: I1206 07:22:30.953857 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:22:30 crc kubenswrapper[4945]: E1206 07:22:30.954694 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:22:42 crc kubenswrapper[4945]: I1206 07:22:42.953338 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:22:42 crc kubenswrapper[4945]: E1206 07:22:42.954084 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:22:56 crc kubenswrapper[4945]: I1206 07:22:56.956465 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:22:56 crc kubenswrapper[4945]: E1206 07:22:56.957228 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:23:10 crc kubenswrapper[4945]: I1206 07:23:10.952993 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:23:10 crc kubenswrapper[4945]: E1206 07:23:10.953911 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.479506 4945 scope.go:117] "RemoveContainer" containerID="3b3520eb06fc60e055fc656ef4c924e2b580496435b938b5e0f0c08ea8880aec" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.531877 4945 scope.go:117] "RemoveContainer" containerID="54e516865e0c2bf291e50817a1f0a0df2b67a1322700ab4cf577d2385c08bcb2" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.566074 4945 scope.go:117] "RemoveContainer" containerID="949300175b977c4e17d92491ddd4ce647b0d87ea1ca4f0ff07636fdbe707cbcc" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.592236 4945 scope.go:117] "RemoveContainer" containerID="de48e720232535cebe13a20fa073ca681801633b91ea03746596672748416839" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.608034 4945 scope.go:117] "RemoveContainer" containerID="a55d092cd1e6de0a0878b76ed91d8a257d77965c701ebff42dc572a4081e9260" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.652727 4945 scope.go:117] "RemoveContainer" containerID="fa3756ffcf3445aad0a9657ca299b426c918cc8325f6302eb842204e4e1fb37a" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.705380 4945 scope.go:117] "RemoveContainer" containerID="fdee576b16798f10cbf0d6df65b22dc489c5de6fce0d46d09574de5ceaf78ca7" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.728825 4945 scope.go:117] "RemoveContainer" containerID="3df7ad5bd72c70655a8062719a5e2173ccd640642cd65ada9584c007120552b2" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.769945 4945 scope.go:117] "RemoveContainer" containerID="4c7e4b215beb1d463e0865db5d67c30bce7decd535274ced24856b5651616753" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.786766 4945 scope.go:117] "RemoveContainer" containerID="2ddc90f531ce8123229116bd612f56498d8293ed1cb1cd9acab8b626c6700c6c" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.804391 4945 scope.go:117] "RemoveContainer" containerID="53c0075dccd3a254fb9efe790919572c497272e3bf93c459ad8b10ca8e292557" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.823235 4945 scope.go:117] "RemoveContainer" containerID="0ec1ad2614ffcda8861fab5a057374555db223dbd1a011d19ff96df9efa454d2" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.855071 4945 scope.go:117] "RemoveContainer" containerID="b56304c17f8860a7fb9704af82c3d821b0358e7f423dc2d3697c62534d14a23a" Dec 06 07:23:15 crc kubenswrapper[4945]: I1206 07:23:15.881168 4945 scope.go:117] "RemoveContainer" containerID="51c9dc30540f672f0cad19373fdd58c9f35dc611b2c4f98ab2d185612f7a7af6" Dec 06 07:23:22 crc kubenswrapper[4945]: I1206 07:23:22.954030 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:23:22 crc kubenswrapper[4945]: E1206 07:23:22.954791 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.285181 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b9d4x"] Dec 06 07:23:33 crc kubenswrapper[4945]: E1206 07:23:33.285996 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ca9fe2-4986-4e85-838b-172332c58972" containerName="registry-server" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.286010 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ca9fe2-4986-4e85-838b-172332c58972" containerName="registry-server" Dec 06 07:23:33 crc kubenswrapper[4945]: E1206 07:23:33.286030 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ca9fe2-4986-4e85-838b-172332c58972" containerName="extract-content" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.286035 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ca9fe2-4986-4e85-838b-172332c58972" containerName="extract-content" Dec 06 07:23:33 crc kubenswrapper[4945]: E1206 07:23:33.286051 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ca9fe2-4986-4e85-838b-172332c58972" containerName="extract-utilities" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.286058 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ca9fe2-4986-4e85-838b-172332c58972" containerName="extract-utilities" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.286199 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ca9fe2-4986-4e85-838b-172332c58972" containerName="registry-server" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.287252 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.303617 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b9d4x"] Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.340623 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdvfd\" (UniqueName: \"kubernetes.io/projected/91fa76e9-3224-4979-b571-67f438cc2128-kube-api-access-tdvfd\") pod \"certified-operators-b9d4x\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.340674 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-catalog-content\") pod \"certified-operators-b9d4x\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.340710 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-utilities\") pod \"certified-operators-b9d4x\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.441810 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdvfd\" (UniqueName: \"kubernetes.io/projected/91fa76e9-3224-4979-b571-67f438cc2128-kube-api-access-tdvfd\") pod \"certified-operators-b9d4x\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.441863 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-catalog-content\") pod \"certified-operators-b9d4x\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.441909 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-utilities\") pod \"certified-operators-b9d4x\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.442470 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-catalog-content\") pod \"certified-operators-b9d4x\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.442524 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-utilities\") pod \"certified-operators-b9d4x\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.471618 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdvfd\" (UniqueName: \"kubernetes.io/projected/91fa76e9-3224-4979-b571-67f438cc2128-kube-api-access-tdvfd\") pod \"certified-operators-b9d4x\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:33 crc kubenswrapper[4945]: I1206 07:23:33.616180 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:34 crc kubenswrapper[4945]: I1206 07:23:34.065516 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b9d4x"] Dec 06 07:23:34 crc kubenswrapper[4945]: I1206 07:23:34.178329 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9d4x" event={"ID":"91fa76e9-3224-4979-b571-67f438cc2128","Type":"ContainerStarted","Data":"8f0daee49466bdc29e52702537541c1faa047503e3f91e3affb6a5f7561fd14f"} Dec 06 07:23:34 crc kubenswrapper[4945]: I1206 07:23:34.954927 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:23:34 crc kubenswrapper[4945]: E1206 07:23:34.955459 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:23:35 crc kubenswrapper[4945]: I1206 07:23:35.190449 4945 generic.go:334] "Generic (PLEG): container finished" podID="91fa76e9-3224-4979-b571-67f438cc2128" containerID="5604943cdb7571f6c6f4b2a48a8efab7723fa7b1ea44cb861f265370ab198ce8" exitCode=0 Dec 06 07:23:35 crc kubenswrapper[4945]: I1206 07:23:35.190492 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9d4x" event={"ID":"91fa76e9-3224-4979-b571-67f438cc2128","Type":"ContainerDied","Data":"5604943cdb7571f6c6f4b2a48a8efab7723fa7b1ea44cb861f265370ab198ce8"} Dec 06 07:23:35 crc kubenswrapper[4945]: I1206 07:23:35.192930 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:23:36 crc kubenswrapper[4945]: I1206 07:23:36.201617 4945 generic.go:334] "Generic (PLEG): container finished" podID="91fa76e9-3224-4979-b571-67f438cc2128" containerID="ef5c480c933c4475943c63cbf8039e06144e1960ac4359f159a93a0cf0a3847b" exitCode=0 Dec 06 07:23:36 crc kubenswrapper[4945]: I1206 07:23:36.201690 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9d4x" event={"ID":"91fa76e9-3224-4979-b571-67f438cc2128","Type":"ContainerDied","Data":"ef5c480c933c4475943c63cbf8039e06144e1960ac4359f159a93a0cf0a3847b"} Dec 06 07:23:37 crc kubenswrapper[4945]: I1206 07:23:37.212533 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9d4x" event={"ID":"91fa76e9-3224-4979-b571-67f438cc2128","Type":"ContainerStarted","Data":"0a8972fb83dc500c8959c561b1cce74621010f298a88d569885d68adb4d49a92"} Dec 06 07:23:37 crc kubenswrapper[4945]: I1206 07:23:37.228694 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b9d4x" podStartSLOduration=2.823650883 podStartE2EDuration="4.228675673s" podCreationTimestamp="2025-12-06 07:23:33 +0000 UTC" firstStartedPulling="2025-12-06 07:23:35.19274769 +0000 UTC m=+1828.647608734" lastFinishedPulling="2025-12-06 07:23:36.59777247 +0000 UTC m=+1830.052633524" observedRunningTime="2025-12-06 07:23:37.226576866 +0000 UTC m=+1830.681437910" watchObservedRunningTime="2025-12-06 07:23:37.228675673 +0000 UTC m=+1830.683536717" Dec 06 07:23:43 crc kubenswrapper[4945]: I1206 07:23:43.617469 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:43 crc kubenswrapper[4945]: I1206 07:23:43.618373 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:43 crc kubenswrapper[4945]: I1206 07:23:43.687771 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:44 crc kubenswrapper[4945]: I1206 07:23:44.314274 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:44 crc kubenswrapper[4945]: I1206 07:23:44.373924 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b9d4x"] Dec 06 07:23:46 crc kubenswrapper[4945]: I1206 07:23:46.286502 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b9d4x" podUID="91fa76e9-3224-4979-b571-67f438cc2128" containerName="registry-server" containerID="cri-o://0a8972fb83dc500c8959c561b1cce74621010f298a88d569885d68adb4d49a92" gracePeriod=2 Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.298941 4945 generic.go:334] "Generic (PLEG): container finished" podID="91fa76e9-3224-4979-b571-67f438cc2128" containerID="0a8972fb83dc500c8959c561b1cce74621010f298a88d569885d68adb4d49a92" exitCode=0 Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.298988 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9d4x" event={"ID":"91fa76e9-3224-4979-b571-67f438cc2128","Type":"ContainerDied","Data":"0a8972fb83dc500c8959c561b1cce74621010f298a88d569885d68adb4d49a92"} Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.795530 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.851076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdvfd\" (UniqueName: \"kubernetes.io/projected/91fa76e9-3224-4979-b571-67f438cc2128-kube-api-access-tdvfd\") pod \"91fa76e9-3224-4979-b571-67f438cc2128\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.851131 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-catalog-content\") pod \"91fa76e9-3224-4979-b571-67f438cc2128\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.851192 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-utilities\") pod \"91fa76e9-3224-4979-b571-67f438cc2128\" (UID: \"91fa76e9-3224-4979-b571-67f438cc2128\") " Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.852341 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-utilities" (OuterVolumeSpecName: "utilities") pod "91fa76e9-3224-4979-b571-67f438cc2128" (UID: "91fa76e9-3224-4979-b571-67f438cc2128"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.857221 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91fa76e9-3224-4979-b571-67f438cc2128-kube-api-access-tdvfd" (OuterVolumeSpecName: "kube-api-access-tdvfd") pod "91fa76e9-3224-4979-b571-67f438cc2128" (UID: "91fa76e9-3224-4979-b571-67f438cc2128"). InnerVolumeSpecName "kube-api-access-tdvfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.910174 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91fa76e9-3224-4979-b571-67f438cc2128" (UID: "91fa76e9-3224-4979-b571-67f438cc2128"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.952217 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdvfd\" (UniqueName: \"kubernetes.io/projected/91fa76e9-3224-4979-b571-67f438cc2128-kube-api-access-tdvfd\") on node \"crc\" DevicePath \"\"" Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.952248 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:23:47 crc kubenswrapper[4945]: I1206 07:23:47.952260 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91fa76e9-3224-4979-b571-67f438cc2128-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:23:48 crc kubenswrapper[4945]: I1206 07:23:48.314772 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9d4x" event={"ID":"91fa76e9-3224-4979-b571-67f438cc2128","Type":"ContainerDied","Data":"8f0daee49466bdc29e52702537541c1faa047503e3f91e3affb6a5f7561fd14f"} Dec 06 07:23:48 crc kubenswrapper[4945]: I1206 07:23:48.314832 4945 scope.go:117] "RemoveContainer" containerID="0a8972fb83dc500c8959c561b1cce74621010f298a88d569885d68adb4d49a92" Dec 06 07:23:48 crc kubenswrapper[4945]: I1206 07:23:48.314970 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9d4x" Dec 06 07:23:48 crc kubenswrapper[4945]: I1206 07:23:48.348514 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b9d4x"] Dec 06 07:23:48 crc kubenswrapper[4945]: I1206 07:23:48.352110 4945 scope.go:117] "RemoveContainer" containerID="ef5c480c933c4475943c63cbf8039e06144e1960ac4359f159a93a0cf0a3847b" Dec 06 07:23:48 crc kubenswrapper[4945]: I1206 07:23:48.358082 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b9d4x"] Dec 06 07:23:48 crc kubenswrapper[4945]: I1206 07:23:48.378049 4945 scope.go:117] "RemoveContainer" containerID="5604943cdb7571f6c6f4b2a48a8efab7723fa7b1ea44cb861f265370ab198ce8" Dec 06 07:23:48 crc kubenswrapper[4945]: I1206 07:23:48.963797 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91fa76e9-3224-4979-b571-67f438cc2128" path="/var/lib/kubelet/pods/91fa76e9-3224-4979-b571-67f438cc2128/volumes" Dec 06 07:23:49 crc kubenswrapper[4945]: I1206 07:23:49.952973 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:23:49 crc kubenswrapper[4945]: E1206 07:23:49.953193 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:24:02 crc kubenswrapper[4945]: I1206 07:24:02.954486 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:24:02 crc kubenswrapper[4945]: E1206 07:24:02.957745 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:24:13 crc kubenswrapper[4945]: I1206 07:24:13.953496 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:24:13 crc kubenswrapper[4945]: E1206 07:24:13.954738 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.046811 4945 scope.go:117] "RemoveContainer" containerID="1041e9cc315c9c6b1e1b5a4a949b745e06f5a68d6531286428930d1c1eb3b390" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.093935 4945 scope.go:117] "RemoveContainer" containerID="c997070637d077fb0bd2b5c6e603ae66c9789fb87271122eac47abaf2b27318e" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.129902 4945 scope.go:117] "RemoveContainer" containerID="8f06adda9c4f122a20f297ca0d7ccd2c0349a987c76bc355daf2b4ea17945592" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.159469 4945 scope.go:117] "RemoveContainer" containerID="87cd0b4a674d5d3593011da803e898cb217a594e41b955bcb125ea3f0c1b9ff8" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.199887 4945 scope.go:117] "RemoveContainer" containerID="36fc5adc9c35a27308a897b987a0ead13d775ca1f0ea63ba9ab9e64e112df411" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.243063 4945 scope.go:117] "RemoveContainer" containerID="f7492d5e5eac181374232bb84375c7fd72e56b1450784144d7ba040bdd005f45" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.268430 4945 scope.go:117] "RemoveContainer" containerID="c13f81a9e25d94e04593bed44e1a89fc9b9c9767b12d18a9ce9a454c2b1fb953" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.305477 4945 scope.go:117] "RemoveContainer" containerID="79de2a8c48d8cdf9a756bedc61a029ecea66e0808f3605f405bca9e9052a6abf" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.342219 4945 scope.go:117] "RemoveContainer" containerID="7cb7cbad5c146d1fa5c2816193c03ecbbc2c2e83f7d5dbab3b49095c873c3919" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.379221 4945 scope.go:117] "RemoveContainer" containerID="df6d926343b2c56b8535b92675f93512dbdb6c74f7bdf758728c976bec5fd4ce" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.415515 4945 scope.go:117] "RemoveContainer" containerID="14269f510b2f262408ff083c20bf00b7f68b48a2e3c497d57f6d2ea6e8cc4fcb" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.440265 4945 scope.go:117] "RemoveContainer" containerID="b957bb81cfc39bf3a8747b1e1804ba1ec5dcb33d0889082d694cf4ffd810ef17" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.469692 4945 scope.go:117] "RemoveContainer" containerID="ed34589057817ad72a5317eab6b9f21905263e7d0873427981949caa82987191" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.498223 4945 scope.go:117] "RemoveContainer" containerID="2326a98863fbfe1d3fcf0bf3e1db275908b04d5165d45b924f0b985263d8c616" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.521208 4945 scope.go:117] "RemoveContainer" containerID="ab370550567ae529abd1c5be85f05c4c95e05cb15eb9877d9089c7f9cdaec299" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.540581 4945 scope.go:117] "RemoveContainer" containerID="acd8fb52b675c118e310b60c8ee78c1ae8f429af506bdaf70f33be2b2e7d0f4d" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.596858 4945 scope.go:117] "RemoveContainer" containerID="c3b041158b4f4c593ae33438ee22c67b72a78b4734276f3449241b8f0221a4dd" Dec 06 07:24:16 crc kubenswrapper[4945]: I1206 07:24:16.619772 4945 scope.go:117] "RemoveContainer" containerID="61c8d1dfdac1f1aac44c2ab0b8448e2373849c62730037840a5d71971f8623dc" Dec 06 07:24:26 crc kubenswrapper[4945]: I1206 07:24:26.964683 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:24:26 crc kubenswrapper[4945]: E1206 07:24:26.966394 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:24:39 crc kubenswrapper[4945]: I1206 07:24:39.954161 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:24:39 crc kubenswrapper[4945]: E1206 07:24:39.955293 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:24:51 crc kubenswrapper[4945]: I1206 07:24:51.953310 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:24:51 crc kubenswrapper[4945]: E1206 07:24:51.954334 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:25:04 crc kubenswrapper[4945]: I1206 07:25:04.953813 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:25:04 crc kubenswrapper[4945]: E1206 07:25:04.954678 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:25:16 crc kubenswrapper[4945]: I1206 07:25:16.864306 4945 scope.go:117] "RemoveContainer" containerID="724de5f0c3d669660e6698a3c99a4875f3d5e9150ef8166ff6741bfbba23a072" Dec 06 07:25:16 crc kubenswrapper[4945]: I1206 07:25:16.907850 4945 scope.go:117] "RemoveContainer" containerID="0f377c0a02117dc7052bb48a4af37d61f3095dd82d7bde891095afbd858d61ee" Dec 06 07:25:16 crc kubenswrapper[4945]: I1206 07:25:16.945537 4945 scope.go:117] "RemoveContainer" containerID="2f935bfecfa8a7af798c6c8d352ae3b0ed1c935c6dba724d358a31059be6f477" Dec 06 07:25:17 crc kubenswrapper[4945]: I1206 07:25:17.008350 4945 scope.go:117] "RemoveContainer" containerID="0ab1417aa10f14854712f6287bc7ff98a77fec3e4030e9d198a868daeb2bd9f6" Dec 06 07:25:17 crc kubenswrapper[4945]: I1206 07:25:17.038600 4945 scope.go:117] "RemoveContainer" containerID="fd592672f76730e847191d2934d9c9cfb67d21204f6a000e593a960e8df07a85" Dec 06 07:25:19 crc kubenswrapper[4945]: I1206 07:25:19.953685 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:25:19 crc kubenswrapper[4945]: E1206 07:25:19.954091 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:25:32 crc kubenswrapper[4945]: I1206 07:25:32.954131 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:25:32 crc kubenswrapper[4945]: E1206 07:25:32.955975 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:25:44 crc kubenswrapper[4945]: I1206 07:25:44.953867 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:25:44 crc kubenswrapper[4945]: E1206 07:25:44.955026 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:25:58 crc kubenswrapper[4945]: I1206 07:25:58.954834 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:25:58 crc kubenswrapper[4945]: E1206 07:25:58.955982 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:26:12 crc kubenswrapper[4945]: I1206 07:26:12.953771 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:26:14 crc kubenswrapper[4945]: I1206 07:26:14.690695 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"8154f198dfb93ec3af16cb96a5a2d7479f9af26793a8679ee0c37d96956caf5a"} Dec 06 07:26:17 crc kubenswrapper[4945]: I1206 07:26:17.128544 4945 scope.go:117] "RemoveContainer" containerID="55421545d7e8b51ad480e6adea462be55656e1f3e06b69e287fb97a724e2a633" Dec 06 07:26:17 crc kubenswrapper[4945]: I1206 07:26:17.161192 4945 scope.go:117] "RemoveContainer" containerID="5984506921a2b100cf94ec4a7e72d52586d9d11cdd69b262d6ee89622fc5fcf4" Dec 06 07:26:17 crc kubenswrapper[4945]: I1206 07:26:17.218788 4945 scope.go:117] "RemoveContainer" containerID="6d8dfeac270484f3d88b61ccf82778281ff7139a1cecec4590cc125407a09c91" Dec 06 07:26:17 crc kubenswrapper[4945]: I1206 07:26:17.265332 4945 scope.go:117] "RemoveContainer" containerID="cf42f6e43b6eaa14b9d60d2d2162c0cfe2ef17791091ee3c53ee8ba8cb1d8339" Dec 06 07:26:17 crc kubenswrapper[4945]: I1206 07:26:17.285755 4945 scope.go:117] "RemoveContainer" containerID="61ecac1b8319891d087db52d2c643d043fe8240ecef32333a71acb3632492838" Dec 06 07:26:17 crc kubenswrapper[4945]: I1206 07:26:17.314268 4945 scope.go:117] "RemoveContainer" containerID="145d4313977864fb199b0a745f9e48fe66f8d8ec1a67bdd8264719366ba4588e" Dec 06 07:27:17 crc kubenswrapper[4945]: I1206 07:27:17.415937 4945 scope.go:117] "RemoveContainer" containerID="ce020a3a06ca7b55178dff84f8b51d68fe50cb6988fa5275e3943fba1bd14855" Dec 06 07:27:17 crc kubenswrapper[4945]: I1206 07:27:17.451669 4945 scope.go:117] "RemoveContainer" containerID="9182d14993c72dff3cfd16a7cb6486a5a13bfa7bb047b47d0805ca2dbeb0e000" Dec 06 07:27:17 crc kubenswrapper[4945]: I1206 07:27:17.480463 4945 scope.go:117] "RemoveContainer" containerID="ca5920be64a83d38ca04c8eb4934c935f589ac9dbbb41a920be90b450c3ed0c1" Dec 06 07:27:17 crc kubenswrapper[4945]: I1206 07:27:17.510795 4945 scope.go:117] "RemoveContainer" containerID="182ff6178fad99f1d5b03a67029019bd53de2d67c4dfbf7f4e430acffa8b1cce" Dec 06 07:28:38 crc kubenswrapper[4945]: I1206 07:28:38.796242 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:28:38 crc kubenswrapper[4945]: I1206 07:28:38.796810 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:29:08 crc kubenswrapper[4945]: I1206 07:29:08.795715 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:29:08 crc kubenswrapper[4945]: I1206 07:29:08.796262 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:29:38 crc kubenswrapper[4945]: I1206 07:29:38.796337 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:29:38 crc kubenswrapper[4945]: I1206 07:29:38.796914 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:29:38 crc kubenswrapper[4945]: I1206 07:29:38.796957 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:29:38 crc kubenswrapper[4945]: I1206 07:29:38.797554 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8154f198dfb93ec3af16cb96a5a2d7479f9af26793a8679ee0c37d96956caf5a"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:29:38 crc kubenswrapper[4945]: I1206 07:29:38.797607 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://8154f198dfb93ec3af16cb96a5a2d7479f9af26793a8679ee0c37d96956caf5a" gracePeriod=600 Dec 06 07:29:39 crc kubenswrapper[4945]: I1206 07:29:39.715673 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="8154f198dfb93ec3af16cb96a5a2d7479f9af26793a8679ee0c37d96956caf5a" exitCode=0 Dec 06 07:29:39 crc kubenswrapper[4945]: I1206 07:29:39.715744 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"8154f198dfb93ec3af16cb96a5a2d7479f9af26793a8679ee0c37d96956caf5a"} Dec 06 07:29:39 crc kubenswrapper[4945]: I1206 07:29:39.717043 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000"} Dec 06 07:29:39 crc kubenswrapper[4945]: I1206 07:29:39.717126 4945 scope.go:117] "RemoveContainer" containerID="ca5d7df51b8dc8e60c1e5246f249cd500fe21c2631e6dab5b099285e05446aed" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.023121 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c7hnp"] Dec 06 07:29:41 crc kubenswrapper[4945]: E1206 07:29:41.023451 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91fa76e9-3224-4979-b571-67f438cc2128" containerName="extract-utilities" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.023466 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="91fa76e9-3224-4979-b571-67f438cc2128" containerName="extract-utilities" Dec 06 07:29:41 crc kubenswrapper[4945]: E1206 07:29:41.023479 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91fa76e9-3224-4979-b571-67f438cc2128" containerName="extract-content" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.023484 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="91fa76e9-3224-4979-b571-67f438cc2128" containerName="extract-content" Dec 06 07:29:41 crc kubenswrapper[4945]: E1206 07:29:41.023492 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91fa76e9-3224-4979-b571-67f438cc2128" containerName="registry-server" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.023497 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="91fa76e9-3224-4979-b571-67f438cc2128" containerName="registry-server" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.023631 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="91fa76e9-3224-4979-b571-67f438cc2128" containerName="registry-server" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.024670 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.034538 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7hnp"] Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.135197 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-utilities\") pod \"redhat-marketplace-c7hnp\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.135316 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-catalog-content\") pod \"redhat-marketplace-c7hnp\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.135360 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl6wf\" (UniqueName: \"kubernetes.io/projected/2f2d856f-228b-404e-8f7d-342ceb20b604-kube-api-access-vl6wf\") pod \"redhat-marketplace-c7hnp\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.236998 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-catalog-content\") pod \"redhat-marketplace-c7hnp\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.237079 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl6wf\" (UniqueName: \"kubernetes.io/projected/2f2d856f-228b-404e-8f7d-342ceb20b604-kube-api-access-vl6wf\") pod \"redhat-marketplace-c7hnp\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.237123 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-utilities\") pod \"redhat-marketplace-c7hnp\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.237609 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-catalog-content\") pod \"redhat-marketplace-c7hnp\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.237701 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-utilities\") pod \"redhat-marketplace-c7hnp\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.257635 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl6wf\" (UniqueName: \"kubernetes.io/projected/2f2d856f-228b-404e-8f7d-342ceb20b604-kube-api-access-vl6wf\") pod \"redhat-marketplace-c7hnp\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.339036 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:41 crc kubenswrapper[4945]: I1206 07:29:41.795552 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7hnp"] Dec 06 07:29:42 crc kubenswrapper[4945]: I1206 07:29:42.746596 4945 generic.go:334] "Generic (PLEG): container finished" podID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerID="9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c" exitCode=0 Dec 06 07:29:42 crc kubenswrapper[4945]: I1206 07:29:42.746643 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7hnp" event={"ID":"2f2d856f-228b-404e-8f7d-342ceb20b604","Type":"ContainerDied","Data":"9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c"} Dec 06 07:29:42 crc kubenswrapper[4945]: I1206 07:29:42.746673 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7hnp" event={"ID":"2f2d856f-228b-404e-8f7d-342ceb20b604","Type":"ContainerStarted","Data":"7d1f6ebb5ad55ac73dae2e3010ebb66dcefec4b1e2552553b230ffd03650beb4"} Dec 06 07:29:42 crc kubenswrapper[4945]: I1206 07:29:42.750150 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:29:43 crc kubenswrapper[4945]: I1206 07:29:43.756012 4945 generic.go:334] "Generic (PLEG): container finished" podID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerID="9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f" exitCode=0 Dec 06 07:29:43 crc kubenswrapper[4945]: I1206 07:29:43.756097 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7hnp" event={"ID":"2f2d856f-228b-404e-8f7d-342ceb20b604","Type":"ContainerDied","Data":"9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f"} Dec 06 07:29:44 crc kubenswrapper[4945]: I1206 07:29:44.770252 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7hnp" event={"ID":"2f2d856f-228b-404e-8f7d-342ceb20b604","Type":"ContainerStarted","Data":"7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718"} Dec 06 07:29:44 crc kubenswrapper[4945]: I1206 07:29:44.787228 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c7hnp" podStartSLOduration=2.345261044 podStartE2EDuration="3.787211283s" podCreationTimestamp="2025-12-06 07:29:41 +0000 UTC" firstStartedPulling="2025-12-06 07:29:42.749867467 +0000 UTC m=+2196.204728521" lastFinishedPulling="2025-12-06 07:29:44.191817706 +0000 UTC m=+2197.646678760" observedRunningTime="2025-12-06 07:29:44.786897204 +0000 UTC m=+2198.241758298" watchObservedRunningTime="2025-12-06 07:29:44.787211283 +0000 UTC m=+2198.242072327" Dec 06 07:29:51 crc kubenswrapper[4945]: I1206 07:29:51.340307 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:51 crc kubenswrapper[4945]: I1206 07:29:51.341088 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:51 crc kubenswrapper[4945]: I1206 07:29:51.391978 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:51 crc kubenswrapper[4945]: I1206 07:29:51.875634 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:51 crc kubenswrapper[4945]: I1206 07:29:51.933536 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7hnp"] Dec 06 07:29:53 crc kubenswrapper[4945]: I1206 07:29:53.836896 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c7hnp" podUID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerName="registry-server" containerID="cri-o://7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718" gracePeriod=2 Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.794940 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.844565 4945 generic.go:334] "Generic (PLEG): container finished" podID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerID="7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718" exitCode=0 Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.844605 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7hnp" event={"ID":"2f2d856f-228b-404e-8f7d-342ceb20b604","Type":"ContainerDied","Data":"7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718"} Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.844632 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7hnp" event={"ID":"2f2d856f-228b-404e-8f7d-342ceb20b604","Type":"ContainerDied","Data":"7d1f6ebb5ad55ac73dae2e3010ebb66dcefec4b1e2552553b230ffd03650beb4"} Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.844632 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7hnp" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.844660 4945 scope.go:117] "RemoveContainer" containerID="7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.861039 4945 scope.go:117] "RemoveContainer" containerID="9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.881844 4945 scope.go:117] "RemoveContainer" containerID="9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.902418 4945 scope.go:117] "RemoveContainer" containerID="7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718" Dec 06 07:29:54 crc kubenswrapper[4945]: E1206 07:29:54.902833 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718\": container with ID starting with 7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718 not found: ID does not exist" containerID="7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.902881 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718"} err="failed to get container status \"7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718\": rpc error: code = NotFound desc = could not find container \"7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718\": container with ID starting with 7935c1e624b6acbd58f1744fbd8ed99d8be6161910f86bda7af1236d8d8fe718 not found: ID does not exist" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.902912 4945 scope.go:117] "RemoveContainer" containerID="9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f" Dec 06 07:29:54 crc kubenswrapper[4945]: E1206 07:29:54.903231 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f\": container with ID starting with 9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f not found: ID does not exist" containerID="9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.903268 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f"} err="failed to get container status \"9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f\": rpc error: code = NotFound desc = could not find container \"9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f\": container with ID starting with 9a677c004e754a6a9080bad1134aa1c6e7d5bdd45c4df711d72d2ab7785bf10f not found: ID does not exist" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.903317 4945 scope.go:117] "RemoveContainer" containerID="9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c" Dec 06 07:29:54 crc kubenswrapper[4945]: E1206 07:29:54.903845 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c\": container with ID starting with 9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c not found: ID does not exist" containerID="9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.903873 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c"} err="failed to get container status \"9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c\": rpc error: code = NotFound desc = could not find container \"9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c\": container with ID starting with 9ba9509cbbde1d9c324b0ce81afef3570da77b6005ba01e2effea73f145bc40c not found: ID does not exist" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.936184 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl6wf\" (UniqueName: \"kubernetes.io/projected/2f2d856f-228b-404e-8f7d-342ceb20b604-kube-api-access-vl6wf\") pod \"2f2d856f-228b-404e-8f7d-342ceb20b604\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.936245 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-utilities\") pod \"2f2d856f-228b-404e-8f7d-342ceb20b604\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.936268 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-catalog-content\") pod \"2f2d856f-228b-404e-8f7d-342ceb20b604\" (UID: \"2f2d856f-228b-404e-8f7d-342ceb20b604\") " Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.937377 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-utilities" (OuterVolumeSpecName: "utilities") pod "2f2d856f-228b-404e-8f7d-342ceb20b604" (UID: "2f2d856f-228b-404e-8f7d-342ceb20b604"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.944399 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f2d856f-228b-404e-8f7d-342ceb20b604-kube-api-access-vl6wf" (OuterVolumeSpecName: "kube-api-access-vl6wf") pod "2f2d856f-228b-404e-8f7d-342ceb20b604" (UID: "2f2d856f-228b-404e-8f7d-342ceb20b604"). InnerVolumeSpecName "kube-api-access-vl6wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:29:54 crc kubenswrapper[4945]: I1206 07:29:54.960803 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f2d856f-228b-404e-8f7d-342ceb20b604" (UID: "2f2d856f-228b-404e-8f7d-342ceb20b604"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:29:55 crc kubenswrapper[4945]: I1206 07:29:55.038662 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl6wf\" (UniqueName: \"kubernetes.io/projected/2f2d856f-228b-404e-8f7d-342ceb20b604-kube-api-access-vl6wf\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:55 crc kubenswrapper[4945]: I1206 07:29:55.038724 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:55 crc kubenswrapper[4945]: I1206 07:29:55.038737 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2d856f-228b-404e-8f7d-342ceb20b604-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:55 crc kubenswrapper[4945]: I1206 07:29:55.171343 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7hnp"] Dec 06 07:29:55 crc kubenswrapper[4945]: I1206 07:29:55.176968 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7hnp"] Dec 06 07:29:56 crc kubenswrapper[4945]: I1206 07:29:56.967414 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f2d856f-228b-404e-8f7d-342ceb20b604" path="/var/lib/kubelet/pods/2f2d856f-228b-404e-8f7d-342ceb20b604/volumes" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.160383 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q"] Dec 06 07:30:00 crc kubenswrapper[4945]: E1206 07:30:00.161033 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerName="extract-utilities" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.161077 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerName="extract-utilities" Dec 06 07:30:00 crc kubenswrapper[4945]: E1206 07:30:00.161110 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerName="extract-content" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.161125 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerName="extract-content" Dec 06 07:30:00 crc kubenswrapper[4945]: E1206 07:30:00.161180 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerName="registry-server" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.161196 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerName="registry-server" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.161537 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f2d856f-228b-404e-8f7d-342ceb20b604" containerName="registry-server" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.162614 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.168340 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.169429 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.171757 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q"] Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.234044 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h8kzt"] Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.236518 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.242843 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h8kzt"] Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.337836 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzdrf\" (UniqueName: \"kubernetes.io/projected/2f287003-14b7-4397-b93f-ffc3a87db524-kube-api-access-hzdrf\") pod \"collect-profiles-29416770-g8n9q\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.338133 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e6adf9-3142-4984-8ee3-02ab261667e6-utilities\") pod \"community-operators-h8kzt\" (UID: \"09e6adf9-3142-4984-8ee3-02ab261667e6\") " pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.338248 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f287003-14b7-4397-b93f-ffc3a87db524-config-volume\") pod \"collect-profiles-29416770-g8n9q\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.338405 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f287003-14b7-4397-b93f-ffc3a87db524-secret-volume\") pod \"collect-profiles-29416770-g8n9q\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.338497 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9jwh\" (UniqueName: \"kubernetes.io/projected/09e6adf9-3142-4984-8ee3-02ab261667e6-kube-api-access-d9jwh\") pod \"community-operators-h8kzt\" (UID: \"09e6adf9-3142-4984-8ee3-02ab261667e6\") " pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.338591 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e6adf9-3142-4984-8ee3-02ab261667e6-catalog-content\") pod \"community-operators-h8kzt\" (UID: \"09e6adf9-3142-4984-8ee3-02ab261667e6\") " pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.440014 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f287003-14b7-4397-b93f-ffc3a87db524-config-volume\") pod \"collect-profiles-29416770-g8n9q\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.440071 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f287003-14b7-4397-b93f-ffc3a87db524-secret-volume\") pod \"collect-profiles-29416770-g8n9q\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.440102 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9jwh\" (UniqueName: \"kubernetes.io/projected/09e6adf9-3142-4984-8ee3-02ab261667e6-kube-api-access-d9jwh\") pod \"community-operators-h8kzt\" (UID: \"09e6adf9-3142-4984-8ee3-02ab261667e6\") " pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.440167 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e6adf9-3142-4984-8ee3-02ab261667e6-catalog-content\") pod \"community-operators-h8kzt\" (UID: \"09e6adf9-3142-4984-8ee3-02ab261667e6\") " pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.440204 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzdrf\" (UniqueName: \"kubernetes.io/projected/2f287003-14b7-4397-b93f-ffc3a87db524-kube-api-access-hzdrf\") pod \"collect-profiles-29416770-g8n9q\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.440231 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e6adf9-3142-4984-8ee3-02ab261667e6-utilities\") pod \"community-operators-h8kzt\" (UID: \"09e6adf9-3142-4984-8ee3-02ab261667e6\") " pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.440823 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e6adf9-3142-4984-8ee3-02ab261667e6-utilities\") pod \"community-operators-h8kzt\" (UID: \"09e6adf9-3142-4984-8ee3-02ab261667e6\") " pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.441037 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f287003-14b7-4397-b93f-ffc3a87db524-config-volume\") pod \"collect-profiles-29416770-g8n9q\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.441101 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e6adf9-3142-4984-8ee3-02ab261667e6-catalog-content\") pod \"community-operators-h8kzt\" (UID: \"09e6adf9-3142-4984-8ee3-02ab261667e6\") " pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.450153 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f287003-14b7-4397-b93f-ffc3a87db524-secret-volume\") pod \"collect-profiles-29416770-g8n9q\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.455754 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzdrf\" (UniqueName: \"kubernetes.io/projected/2f287003-14b7-4397-b93f-ffc3a87db524-kube-api-access-hzdrf\") pod \"collect-profiles-29416770-g8n9q\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.458187 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9jwh\" (UniqueName: \"kubernetes.io/projected/09e6adf9-3142-4984-8ee3-02ab261667e6-kube-api-access-d9jwh\") pod \"community-operators-h8kzt\" (UID: \"09e6adf9-3142-4984-8ee3-02ab261667e6\") " pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.483345 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.560052 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:00 crc kubenswrapper[4945]: I1206 07:30:00.940270 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q"] Dec 06 07:30:01 crc kubenswrapper[4945]: I1206 07:30:01.088627 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h8kzt"] Dec 06 07:30:01 crc kubenswrapper[4945]: I1206 07:30:01.935419 4945 generic.go:334] "Generic (PLEG): container finished" podID="2f287003-14b7-4397-b93f-ffc3a87db524" containerID="3cf427c5a6ae004a4d62091b57effad841cd9e1e41fdee962577561e499dbeb5" exitCode=0 Dec 06 07:30:01 crc kubenswrapper[4945]: I1206 07:30:01.935576 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" event={"ID":"2f287003-14b7-4397-b93f-ffc3a87db524","Type":"ContainerDied","Data":"3cf427c5a6ae004a4d62091b57effad841cd9e1e41fdee962577561e499dbeb5"} Dec 06 07:30:01 crc kubenswrapper[4945]: I1206 07:30:01.935763 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" event={"ID":"2f287003-14b7-4397-b93f-ffc3a87db524","Type":"ContainerStarted","Data":"241b3889000b509eb81f4c7249fc7a085a5d1734d919361ae1b469b540b4429a"} Dec 06 07:30:01 crc kubenswrapper[4945]: I1206 07:30:01.939124 4945 generic.go:334] "Generic (PLEG): container finished" podID="09e6adf9-3142-4984-8ee3-02ab261667e6" containerID="3ea69d96ad13a49bae5ecfac805c357286f4c3cad080d894bc7fdb208c914b41" exitCode=0 Dec 06 07:30:01 crc kubenswrapper[4945]: I1206 07:30:01.939172 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8kzt" event={"ID":"09e6adf9-3142-4984-8ee3-02ab261667e6","Type":"ContainerDied","Data":"3ea69d96ad13a49bae5ecfac805c357286f4c3cad080d894bc7fdb208c914b41"} Dec 06 07:30:01 crc kubenswrapper[4945]: I1206 07:30:01.939205 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8kzt" event={"ID":"09e6adf9-3142-4984-8ee3-02ab261667e6","Type":"ContainerStarted","Data":"a8bd0623b4449b1ec1810af40eef021cdf04db2d1d75d3f08b96296e361f2f39"} Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.235590 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.388487 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f287003-14b7-4397-b93f-ffc3a87db524-config-volume\") pod \"2f287003-14b7-4397-b93f-ffc3a87db524\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.388736 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzdrf\" (UniqueName: \"kubernetes.io/projected/2f287003-14b7-4397-b93f-ffc3a87db524-kube-api-access-hzdrf\") pod \"2f287003-14b7-4397-b93f-ffc3a87db524\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.388847 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f287003-14b7-4397-b93f-ffc3a87db524-secret-volume\") pod \"2f287003-14b7-4397-b93f-ffc3a87db524\" (UID: \"2f287003-14b7-4397-b93f-ffc3a87db524\") " Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.390250 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f287003-14b7-4397-b93f-ffc3a87db524-config-volume" (OuterVolumeSpecName: "config-volume") pod "2f287003-14b7-4397-b93f-ffc3a87db524" (UID: "2f287003-14b7-4397-b93f-ffc3a87db524"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.396194 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f287003-14b7-4397-b93f-ffc3a87db524-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2f287003-14b7-4397-b93f-ffc3a87db524" (UID: "2f287003-14b7-4397-b93f-ffc3a87db524"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.397175 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f287003-14b7-4397-b93f-ffc3a87db524-kube-api-access-hzdrf" (OuterVolumeSpecName: "kube-api-access-hzdrf") pod "2f287003-14b7-4397-b93f-ffc3a87db524" (UID: "2f287003-14b7-4397-b93f-ffc3a87db524"). InnerVolumeSpecName "kube-api-access-hzdrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.490589 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f287003-14b7-4397-b93f-ffc3a87db524-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.490629 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f287003-14b7-4397-b93f-ffc3a87db524-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.490642 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzdrf\" (UniqueName: \"kubernetes.io/projected/2f287003-14b7-4397-b93f-ffc3a87db524-kube-api-access-hzdrf\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.957645 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" event={"ID":"2f287003-14b7-4397-b93f-ffc3a87db524","Type":"ContainerDied","Data":"241b3889000b509eb81f4c7249fc7a085a5d1734d919361ae1b469b540b4429a"} Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.957692 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="241b3889000b509eb81f4c7249fc7a085a5d1734d919361ae1b469b540b4429a" Dec 06 07:30:03 crc kubenswrapper[4945]: I1206 07:30:03.957711 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q" Dec 06 07:30:04 crc kubenswrapper[4945]: I1206 07:30:04.315320 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf"] Dec 06 07:30:04 crc kubenswrapper[4945]: I1206 07:30:04.320912 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-r5wbf"] Dec 06 07:30:04 crc kubenswrapper[4945]: I1206 07:30:04.988220 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a24d873-f77d-43d9-a402-b7a3d9194496" path="/var/lib/kubelet/pods/9a24d873-f77d-43d9-a402-b7a3d9194496/volumes" Dec 06 07:30:05 crc kubenswrapper[4945]: I1206 07:30:05.990156 4945 generic.go:334] "Generic (PLEG): container finished" podID="09e6adf9-3142-4984-8ee3-02ab261667e6" containerID="5d12c3bed569e91f548a47b3b9fc38a3b7428e032839f615fafa53e689acad1d" exitCode=0 Dec 06 07:30:05 crc kubenswrapper[4945]: I1206 07:30:05.990240 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8kzt" event={"ID":"09e6adf9-3142-4984-8ee3-02ab261667e6","Type":"ContainerDied","Data":"5d12c3bed569e91f548a47b3b9fc38a3b7428e032839f615fafa53e689acad1d"} Dec 06 07:30:07 crc kubenswrapper[4945]: I1206 07:30:07.001965 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8kzt" event={"ID":"09e6adf9-3142-4984-8ee3-02ab261667e6","Type":"ContainerStarted","Data":"96bc2e2cfab912b80021da4be81dcb7fca53e2e5a606c10ed18e3e47def35bf6"} Dec 06 07:30:10 crc kubenswrapper[4945]: I1206 07:30:10.560657 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:10 crc kubenswrapper[4945]: I1206 07:30:10.561037 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:10 crc kubenswrapper[4945]: I1206 07:30:10.640111 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:10 crc kubenswrapper[4945]: I1206 07:30:10.676750 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h8kzt" podStartSLOduration=6.156788636 podStartE2EDuration="10.67672476s" podCreationTimestamp="2025-12-06 07:30:00 +0000 UTC" firstStartedPulling="2025-12-06 07:30:01.943041038 +0000 UTC m=+2215.397902112" lastFinishedPulling="2025-12-06 07:30:06.462977192 +0000 UTC m=+2219.917838236" observedRunningTime="2025-12-06 07:30:07.032544244 +0000 UTC m=+2220.487405278" watchObservedRunningTime="2025-12-06 07:30:10.67672476 +0000 UTC m=+2224.131585844" Dec 06 07:30:11 crc kubenswrapper[4945]: I1206 07:30:11.081379 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h8kzt" Dec 06 07:30:11 crc kubenswrapper[4945]: I1206 07:30:11.720220 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h8kzt"] Dec 06 07:30:11 crc kubenswrapper[4945]: I1206 07:30:11.897947 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n9zht"] Dec 06 07:30:11 crc kubenswrapper[4945]: I1206 07:30:11.898439 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n9zht" podUID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerName="registry-server" containerID="cri-o://163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186" gracePeriod=2 Dec 06 07:30:12 crc kubenswrapper[4945]: I1206 07:30:12.783176 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9zht" Dec 06 07:30:12 crc kubenswrapper[4945]: I1206 07:30:12.923517 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-utilities\") pod \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " Dec 06 07:30:12 crc kubenswrapper[4945]: I1206 07:30:12.923625 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-catalog-content\") pod \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " Dec 06 07:30:12 crc kubenswrapper[4945]: I1206 07:30:12.923722 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj75j\" (UniqueName: \"kubernetes.io/projected/17d087a6-7ecb-4b21-b825-49fd80f6faf4-kube-api-access-kj75j\") pod \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\" (UID: \"17d087a6-7ecb-4b21-b825-49fd80f6faf4\") " Dec 06 07:30:12 crc kubenswrapper[4945]: I1206 07:30:12.924167 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-utilities" (OuterVolumeSpecName: "utilities") pod "17d087a6-7ecb-4b21-b825-49fd80f6faf4" (UID: "17d087a6-7ecb-4b21-b825-49fd80f6faf4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:30:12 crc kubenswrapper[4945]: I1206 07:30:12.933133 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d087a6-7ecb-4b21-b825-49fd80f6faf4-kube-api-access-kj75j" (OuterVolumeSpecName: "kube-api-access-kj75j") pod "17d087a6-7ecb-4b21-b825-49fd80f6faf4" (UID: "17d087a6-7ecb-4b21-b825-49fd80f6faf4"). InnerVolumeSpecName "kube-api-access-kj75j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:30:12 crc kubenswrapper[4945]: I1206 07:30:12.974419 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17d087a6-7ecb-4b21-b825-49fd80f6faf4" (UID: "17d087a6-7ecb-4b21-b825-49fd80f6faf4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.025714 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.025755 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj75j\" (UniqueName: \"kubernetes.io/projected/17d087a6-7ecb-4b21-b825-49fd80f6faf4-kube-api-access-kj75j\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.025770 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d087a6-7ecb-4b21-b825-49fd80f6faf4-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.050198 4945 generic.go:334] "Generic (PLEG): container finished" podID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerID="163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186" exitCode=0 Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.050244 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9zht" event={"ID":"17d087a6-7ecb-4b21-b825-49fd80f6faf4","Type":"ContainerDied","Data":"163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186"} Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.050348 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9zht" event={"ID":"17d087a6-7ecb-4b21-b825-49fd80f6faf4","Type":"ContainerDied","Data":"8a30ec4ffdad40b4a9860ed83d4b3f59d37d72946ded1ad633d1193c0f18868d"} Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.050374 4945 scope.go:117] "RemoveContainer" containerID="163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.050714 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9zht" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.086760 4945 scope.go:117] "RemoveContainer" containerID="ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.098253 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n9zht"] Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.108555 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n9zht"] Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.132693 4945 scope.go:117] "RemoveContainer" containerID="bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.149723 4945 scope.go:117] "RemoveContainer" containerID="163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186" Dec 06 07:30:13 crc kubenswrapper[4945]: E1206 07:30:13.150130 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186\": container with ID starting with 163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186 not found: ID does not exist" containerID="163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.150176 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186"} err="failed to get container status \"163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186\": rpc error: code = NotFound desc = could not find container \"163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186\": container with ID starting with 163cc769aa4482245b4b9a43e19de1636987828530cb609ab2a7e20ff0821186 not found: ID does not exist" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.150202 4945 scope.go:117] "RemoveContainer" containerID="ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea" Dec 06 07:30:13 crc kubenswrapper[4945]: E1206 07:30:13.150740 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea\": container with ID starting with ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea not found: ID does not exist" containerID="ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.150805 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea"} err="failed to get container status \"ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea\": rpc error: code = NotFound desc = could not find container \"ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea\": container with ID starting with ac5ffada76de785c585fdfc0588eb6945ffbc2d61d65e17df1d209e2e77f6eea not found: ID does not exist" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.150838 4945 scope.go:117] "RemoveContainer" containerID="bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9" Dec 06 07:30:13 crc kubenswrapper[4945]: E1206 07:30:13.151161 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9\": container with ID starting with bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9 not found: ID does not exist" containerID="bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9" Dec 06 07:30:13 crc kubenswrapper[4945]: I1206 07:30:13.151193 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9"} err="failed to get container status \"bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9\": rpc error: code = NotFound desc = could not find container \"bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9\": container with ID starting with bd2c0568587e3c6979b7841a7fdd4befa028854f4d87af3b7a9be81dd10f59c9 not found: ID does not exist" Dec 06 07:30:14 crc kubenswrapper[4945]: I1206 07:30:14.963005 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" path="/var/lib/kubelet/pods/17d087a6-7ecb-4b21-b825-49fd80f6faf4/volumes" Dec 06 07:30:17 crc kubenswrapper[4945]: I1206 07:30:17.657891 4945 scope.go:117] "RemoveContainer" containerID="f9c0b76a4d002b08f0e0c15b7b1e4b811f8c94e85cc23e4767ee705f7ac3ec1b" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.652139 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8lsns"] Dec 06 07:32:01 crc kubenswrapper[4945]: E1206 07:32:01.656894 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerName="extract-content" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.656919 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerName="extract-content" Dec 06 07:32:01 crc kubenswrapper[4945]: E1206 07:32:01.656955 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerName="extract-utilities" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.656965 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerName="extract-utilities" Dec 06 07:32:01 crc kubenswrapper[4945]: E1206 07:32:01.656984 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f287003-14b7-4397-b93f-ffc3a87db524" containerName="collect-profiles" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.656995 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f287003-14b7-4397-b93f-ffc3a87db524" containerName="collect-profiles" Dec 06 07:32:01 crc kubenswrapper[4945]: E1206 07:32:01.657006 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerName="registry-server" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.657015 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerName="registry-server" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.657235 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f287003-14b7-4397-b93f-ffc3a87db524" containerName="collect-profiles" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.657263 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d087a6-7ecb-4b21-b825-49fd80f6faf4" containerName="registry-server" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.658884 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.696496 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8lsns"] Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.756434 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-utilities\") pod \"redhat-operators-8lsns\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.756510 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-catalog-content\") pod \"redhat-operators-8lsns\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.756622 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs5d9\" (UniqueName: \"kubernetes.io/projected/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-kube-api-access-cs5d9\") pod \"redhat-operators-8lsns\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.857937 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs5d9\" (UniqueName: \"kubernetes.io/projected/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-kube-api-access-cs5d9\") pod \"redhat-operators-8lsns\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.858706 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-utilities\") pod \"redhat-operators-8lsns\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.858746 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-utilities\") pod \"redhat-operators-8lsns\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.858794 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-catalog-content\") pod \"redhat-operators-8lsns\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.859041 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-catalog-content\") pod \"redhat-operators-8lsns\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.879603 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs5d9\" (UniqueName: \"kubernetes.io/projected/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-kube-api-access-cs5d9\") pod \"redhat-operators-8lsns\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:01 crc kubenswrapper[4945]: I1206 07:32:01.988186 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:02 crc kubenswrapper[4945]: I1206 07:32:02.418255 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8lsns"] Dec 06 07:32:03 crc kubenswrapper[4945]: I1206 07:32:03.098804 4945 generic.go:334] "Generic (PLEG): container finished" podID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerID="01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9" exitCode=0 Dec 06 07:32:03 crc kubenswrapper[4945]: I1206 07:32:03.098871 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8lsns" event={"ID":"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524","Type":"ContainerDied","Data":"01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9"} Dec 06 07:32:03 crc kubenswrapper[4945]: I1206 07:32:03.099055 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8lsns" event={"ID":"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524","Type":"ContainerStarted","Data":"2e9c73c57ba7507132516ff9cb405eca068e671be68402784c3e1d14f4d1e49b"} Dec 06 07:32:04 crc kubenswrapper[4945]: I1206 07:32:04.112711 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8lsns" event={"ID":"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524","Type":"ContainerStarted","Data":"30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132"} Dec 06 07:32:05 crc kubenswrapper[4945]: I1206 07:32:05.125410 4945 generic.go:334] "Generic (PLEG): container finished" podID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerID="30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132" exitCode=0 Dec 06 07:32:05 crc kubenswrapper[4945]: I1206 07:32:05.125474 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8lsns" event={"ID":"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524","Type":"ContainerDied","Data":"30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132"} Dec 06 07:32:06 crc kubenswrapper[4945]: I1206 07:32:06.138686 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8lsns" event={"ID":"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524","Type":"ContainerStarted","Data":"96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d"} Dec 06 07:32:06 crc kubenswrapper[4945]: I1206 07:32:06.167700 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8lsns" podStartSLOduration=2.550322934 podStartE2EDuration="5.167676843s" podCreationTimestamp="2025-12-06 07:32:01 +0000 UTC" firstStartedPulling="2025-12-06 07:32:03.100998222 +0000 UTC m=+2336.555859266" lastFinishedPulling="2025-12-06 07:32:05.718352131 +0000 UTC m=+2339.173213175" observedRunningTime="2025-12-06 07:32:06.163964673 +0000 UTC m=+2339.618825767" watchObservedRunningTime="2025-12-06 07:32:06.167676843 +0000 UTC m=+2339.622537897" Dec 06 07:32:08 crc kubenswrapper[4945]: I1206 07:32:08.795445 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:32:08 crc kubenswrapper[4945]: I1206 07:32:08.795803 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:32:11 crc kubenswrapper[4945]: I1206 07:32:11.988400 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:11 crc kubenswrapper[4945]: I1206 07:32:11.988905 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:13 crc kubenswrapper[4945]: I1206 07:32:13.062643 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8lsns" podUID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerName="registry-server" probeResult="failure" output=< Dec 06 07:32:13 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 07:32:13 crc kubenswrapper[4945]: > Dec 06 07:32:22 crc kubenswrapper[4945]: I1206 07:32:22.067864 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:22 crc kubenswrapper[4945]: I1206 07:32:22.149073 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:22 crc kubenswrapper[4945]: I1206 07:32:22.311615 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8lsns"] Dec 06 07:32:23 crc kubenswrapper[4945]: I1206 07:32:23.302724 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8lsns" podUID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerName="registry-server" containerID="cri-o://96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d" gracePeriod=2 Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.268173 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.319957 4945 generic.go:334] "Generic (PLEG): container finished" podID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerID="96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d" exitCode=0 Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.320009 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8lsns" event={"ID":"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524","Type":"ContainerDied","Data":"96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d"} Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.320043 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8lsns" event={"ID":"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524","Type":"ContainerDied","Data":"2e9c73c57ba7507132516ff9cb405eca068e671be68402784c3e1d14f4d1e49b"} Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.320068 4945 scope.go:117] "RemoveContainer" containerID="96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.320157 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8lsns" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.352851 4945 scope.go:117] "RemoveContainer" containerID="30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.384361 4945 scope.go:117] "RemoveContainer" containerID="01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.412798 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-utilities\") pod \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.412996 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs5d9\" (UniqueName: \"kubernetes.io/projected/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-kube-api-access-cs5d9\") pod \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.413026 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-catalog-content\") pod \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\" (UID: \"621d1ba7-4e6e-4eb1-af7f-5dac67e5f524\") " Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.414816 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-utilities" (OuterVolumeSpecName: "utilities") pod "621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" (UID: "621d1ba7-4e6e-4eb1-af7f-5dac67e5f524"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.423688 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-kube-api-access-cs5d9" (OuterVolumeSpecName: "kube-api-access-cs5d9") pod "621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" (UID: "621d1ba7-4e6e-4eb1-af7f-5dac67e5f524"). InnerVolumeSpecName "kube-api-access-cs5d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.424185 4945 scope.go:117] "RemoveContainer" containerID="96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d" Dec 06 07:32:24 crc kubenswrapper[4945]: E1206 07:32:24.425216 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d\": container with ID starting with 96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d not found: ID does not exist" containerID="96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.425354 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d"} err="failed to get container status \"96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d\": rpc error: code = NotFound desc = could not find container \"96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d\": container with ID starting with 96e4010a6b0c25eb90d998477f2d728fa132dd6c282aa0fbc0c702f4f20cde4d not found: ID does not exist" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.425434 4945 scope.go:117] "RemoveContainer" containerID="30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132" Dec 06 07:32:24 crc kubenswrapper[4945]: E1206 07:32:24.426088 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132\": container with ID starting with 30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132 not found: ID does not exist" containerID="30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.426137 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132"} err="failed to get container status \"30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132\": rpc error: code = NotFound desc = could not find container \"30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132\": container with ID starting with 30b34e012f25ae703daf25eabddca1a7965a5af00db2aa5986ba946464322132 not found: ID does not exist" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.426167 4945 scope.go:117] "RemoveContainer" containerID="01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9" Dec 06 07:32:24 crc kubenswrapper[4945]: E1206 07:32:24.426682 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9\": container with ID starting with 01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9 not found: ID does not exist" containerID="01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.426823 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9"} err="failed to get container status \"01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9\": rpc error: code = NotFound desc = could not find container \"01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9\": container with ID starting with 01c8da2aa801ec0d9a705a3350cc00734b5a36aee6500530c1e33e263b1321a9 not found: ID does not exist" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.515254 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs5d9\" (UniqueName: \"kubernetes.io/projected/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-kube-api-access-cs5d9\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.515376 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.568920 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" (UID: "621d1ba7-4e6e-4eb1-af7f-5dac67e5f524"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.617492 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.688422 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8lsns"] Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.701337 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8lsns"] Dec 06 07:32:24 crc kubenswrapper[4945]: I1206 07:32:24.968202 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" path="/var/lib/kubelet/pods/621d1ba7-4e6e-4eb1-af7f-5dac67e5f524/volumes" Dec 06 07:32:38 crc kubenswrapper[4945]: I1206 07:32:38.795877 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:32:38 crc kubenswrapper[4945]: I1206 07:32:38.796331 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:33:08 crc kubenswrapper[4945]: I1206 07:33:08.795514 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:33:08 crc kubenswrapper[4945]: I1206 07:33:08.796102 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:33:08 crc kubenswrapper[4945]: I1206 07:33:08.796145 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:33:08 crc kubenswrapper[4945]: I1206 07:33:08.796595 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:33:08 crc kubenswrapper[4945]: I1206 07:33:08.796650 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" gracePeriod=600 Dec 06 07:33:08 crc kubenswrapper[4945]: E1206 07:33:08.916080 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:33:09 crc kubenswrapper[4945]: I1206 07:33:09.055840 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" exitCode=0 Dec 06 07:33:09 crc kubenswrapper[4945]: I1206 07:33:09.055899 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000"} Dec 06 07:33:09 crc kubenswrapper[4945]: I1206 07:33:09.055944 4945 scope.go:117] "RemoveContainer" containerID="8154f198dfb93ec3af16cb96a5a2d7479f9af26793a8679ee0c37d96956caf5a" Dec 06 07:33:09 crc kubenswrapper[4945]: I1206 07:33:09.056629 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:33:09 crc kubenswrapper[4945]: E1206 07:33:09.056966 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:33:23 crc kubenswrapper[4945]: I1206 07:33:23.953602 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:33:23 crc kubenswrapper[4945]: E1206 07:33:23.955120 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:33:36 crc kubenswrapper[4945]: I1206 07:33:36.962550 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:33:36 crc kubenswrapper[4945]: E1206 07:33:36.963628 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.319453 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rhq4n"] Dec 06 07:33:43 crc kubenswrapper[4945]: E1206 07:33:43.320360 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerName="registry-server" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.320377 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerName="registry-server" Dec 06 07:33:43 crc kubenswrapper[4945]: E1206 07:33:43.320397 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerName="extract-content" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.320406 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerName="extract-content" Dec 06 07:33:43 crc kubenswrapper[4945]: E1206 07:33:43.320427 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerName="extract-utilities" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.320439 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerName="extract-utilities" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.320614 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="621d1ba7-4e6e-4eb1-af7f-5dac67e5f524" containerName="registry-server" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.322191 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.334560 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rhq4n"] Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.393469 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmlxc\" (UniqueName: \"kubernetes.io/projected/ef133557-52bb-4cc5-bba4-f44efb51a45d-kube-api-access-pmlxc\") pod \"certified-operators-rhq4n\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.393604 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-catalog-content\") pod \"certified-operators-rhq4n\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.393705 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-utilities\") pod \"certified-operators-rhq4n\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.496272 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmlxc\" (UniqueName: \"kubernetes.io/projected/ef133557-52bb-4cc5-bba4-f44efb51a45d-kube-api-access-pmlxc\") pod \"certified-operators-rhq4n\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.496457 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-catalog-content\") pod \"certified-operators-rhq4n\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.496540 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-utilities\") pod \"certified-operators-rhq4n\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.498594 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-utilities\") pod \"certified-operators-rhq4n\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.498645 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-catalog-content\") pod \"certified-operators-rhq4n\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.526191 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmlxc\" (UniqueName: \"kubernetes.io/projected/ef133557-52bb-4cc5-bba4-f44efb51a45d-kube-api-access-pmlxc\") pod \"certified-operators-rhq4n\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:43 crc kubenswrapper[4945]: I1206 07:33:43.657662 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:44 crc kubenswrapper[4945]: I1206 07:33:44.117519 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rhq4n"] Dec 06 07:33:44 crc kubenswrapper[4945]: I1206 07:33:44.344087 4945 generic.go:334] "Generic (PLEG): container finished" podID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerID="4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48" exitCode=0 Dec 06 07:33:44 crc kubenswrapper[4945]: I1206 07:33:44.344146 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhq4n" event={"ID":"ef133557-52bb-4cc5-bba4-f44efb51a45d","Type":"ContainerDied","Data":"4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48"} Dec 06 07:33:44 crc kubenswrapper[4945]: I1206 07:33:44.344183 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhq4n" event={"ID":"ef133557-52bb-4cc5-bba4-f44efb51a45d","Type":"ContainerStarted","Data":"01c11ea0d54863e7cf865995ed50d978b8b325c5004b064adeb219b5c3856bd3"} Dec 06 07:33:45 crc kubenswrapper[4945]: I1206 07:33:45.355073 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhq4n" event={"ID":"ef133557-52bb-4cc5-bba4-f44efb51a45d","Type":"ContainerStarted","Data":"63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756"} Dec 06 07:33:46 crc kubenswrapper[4945]: I1206 07:33:46.367605 4945 generic.go:334] "Generic (PLEG): container finished" podID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerID="63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756" exitCode=0 Dec 06 07:33:46 crc kubenswrapper[4945]: I1206 07:33:46.367728 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhq4n" event={"ID":"ef133557-52bb-4cc5-bba4-f44efb51a45d","Type":"ContainerDied","Data":"63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756"} Dec 06 07:33:47 crc kubenswrapper[4945]: I1206 07:33:47.376332 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhq4n" event={"ID":"ef133557-52bb-4cc5-bba4-f44efb51a45d","Type":"ContainerStarted","Data":"32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73"} Dec 06 07:33:47 crc kubenswrapper[4945]: I1206 07:33:47.397138 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rhq4n" podStartSLOduration=1.878629294 podStartE2EDuration="4.397125171s" podCreationTimestamp="2025-12-06 07:33:43 +0000 UTC" firstStartedPulling="2025-12-06 07:33:44.345569831 +0000 UTC m=+2437.800430875" lastFinishedPulling="2025-12-06 07:33:46.864065708 +0000 UTC m=+2440.318926752" observedRunningTime="2025-12-06 07:33:47.392661511 +0000 UTC m=+2440.847522555" watchObservedRunningTime="2025-12-06 07:33:47.397125171 +0000 UTC m=+2440.851986215" Dec 06 07:33:51 crc kubenswrapper[4945]: I1206 07:33:51.953873 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:33:51 crc kubenswrapper[4945]: E1206 07:33:51.955010 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:33:53 crc kubenswrapper[4945]: I1206 07:33:53.658657 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:53 crc kubenswrapper[4945]: I1206 07:33:53.658957 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:53 crc kubenswrapper[4945]: I1206 07:33:53.730788 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:54 crc kubenswrapper[4945]: I1206 07:33:54.503306 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:54 crc kubenswrapper[4945]: I1206 07:33:54.565764 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rhq4n"] Dec 06 07:33:56 crc kubenswrapper[4945]: I1206 07:33:56.443884 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rhq4n" podUID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerName="registry-server" containerID="cri-o://32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73" gracePeriod=2 Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.435665 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.456092 4945 generic.go:334] "Generic (PLEG): container finished" podID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerID="32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73" exitCode=0 Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.456139 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhq4n" event={"ID":"ef133557-52bb-4cc5-bba4-f44efb51a45d","Type":"ContainerDied","Data":"32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73"} Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.456171 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhq4n" event={"ID":"ef133557-52bb-4cc5-bba4-f44efb51a45d","Type":"ContainerDied","Data":"01c11ea0d54863e7cf865995ed50d978b8b325c5004b064adeb219b5c3856bd3"} Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.456192 4945 scope.go:117] "RemoveContainer" containerID="32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.456362 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhq4n" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.480317 4945 scope.go:117] "RemoveContainer" containerID="63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.504906 4945 scope.go:117] "RemoveContainer" containerID="4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.529640 4945 scope.go:117] "RemoveContainer" containerID="32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73" Dec 06 07:33:57 crc kubenswrapper[4945]: E1206 07:33:57.530375 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73\": container with ID starting with 32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73 not found: ID does not exist" containerID="32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.530425 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73"} err="failed to get container status \"32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73\": rpc error: code = NotFound desc = could not find container \"32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73\": container with ID starting with 32d4858b136a507d6325c7fdd3fc4e24c913759375b94719c8f305de97012a73 not found: ID does not exist" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.530452 4945 scope.go:117] "RemoveContainer" containerID="63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756" Dec 06 07:33:57 crc kubenswrapper[4945]: E1206 07:33:57.530865 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756\": container with ID starting with 63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756 not found: ID does not exist" containerID="63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.530928 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756"} err="failed to get container status \"63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756\": rpc error: code = NotFound desc = could not find container \"63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756\": container with ID starting with 63ec2b99a971ca901f9e3477b783387482b3bed6821d387fa1dd8c6a4c228756 not found: ID does not exist" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.530955 4945 scope.go:117] "RemoveContainer" containerID="4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48" Dec 06 07:33:57 crc kubenswrapper[4945]: E1206 07:33:57.531461 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48\": container with ID starting with 4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48 not found: ID does not exist" containerID="4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.531505 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48"} err="failed to get container status \"4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48\": rpc error: code = NotFound desc = could not find container \"4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48\": container with ID starting with 4f651c8b8d64f71b2ab6725b7cd8b6e306834b0dbabb411271c88964ccf09b48 not found: ID does not exist" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.532869 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmlxc\" (UniqueName: \"kubernetes.io/projected/ef133557-52bb-4cc5-bba4-f44efb51a45d-kube-api-access-pmlxc\") pod \"ef133557-52bb-4cc5-bba4-f44efb51a45d\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.532966 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-utilities\") pod \"ef133557-52bb-4cc5-bba4-f44efb51a45d\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.533030 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-catalog-content\") pod \"ef133557-52bb-4cc5-bba4-f44efb51a45d\" (UID: \"ef133557-52bb-4cc5-bba4-f44efb51a45d\") " Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.534139 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-utilities" (OuterVolumeSpecName: "utilities") pod "ef133557-52bb-4cc5-bba4-f44efb51a45d" (UID: "ef133557-52bb-4cc5-bba4-f44efb51a45d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.538812 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef133557-52bb-4cc5-bba4-f44efb51a45d-kube-api-access-pmlxc" (OuterVolumeSpecName: "kube-api-access-pmlxc") pod "ef133557-52bb-4cc5-bba4-f44efb51a45d" (UID: "ef133557-52bb-4cc5-bba4-f44efb51a45d"). InnerVolumeSpecName "kube-api-access-pmlxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.601562 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef133557-52bb-4cc5-bba4-f44efb51a45d" (UID: "ef133557-52bb-4cc5-bba4-f44efb51a45d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.634710 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmlxc\" (UniqueName: \"kubernetes.io/projected/ef133557-52bb-4cc5-bba4-f44efb51a45d-kube-api-access-pmlxc\") on node \"crc\" DevicePath \"\"" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.634743 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.634757 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef133557-52bb-4cc5-bba4-f44efb51a45d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.796361 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rhq4n"] Dec 06 07:33:57 crc kubenswrapper[4945]: I1206 07:33:57.806776 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rhq4n"] Dec 06 07:33:58 crc kubenswrapper[4945]: I1206 07:33:58.970876 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef133557-52bb-4cc5-bba4-f44efb51a45d" path="/var/lib/kubelet/pods/ef133557-52bb-4cc5-bba4-f44efb51a45d/volumes" Dec 06 07:34:02 crc kubenswrapper[4945]: I1206 07:34:02.953985 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:34:02 crc kubenswrapper[4945]: E1206 07:34:02.955027 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:34:13 crc kubenswrapper[4945]: I1206 07:34:13.954465 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:34:13 crc kubenswrapper[4945]: E1206 07:34:13.955388 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:34:27 crc kubenswrapper[4945]: I1206 07:34:27.953748 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:34:27 crc kubenswrapper[4945]: E1206 07:34:27.955120 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:34:38 crc kubenswrapper[4945]: I1206 07:34:38.953215 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:34:38 crc kubenswrapper[4945]: E1206 07:34:38.954016 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:34:49 crc kubenswrapper[4945]: I1206 07:34:49.953878 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:34:49 crc kubenswrapper[4945]: E1206 07:34:49.954667 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:35:00 crc kubenswrapper[4945]: I1206 07:35:00.953617 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:35:00 crc kubenswrapper[4945]: E1206 07:35:00.954485 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:35:13 crc kubenswrapper[4945]: I1206 07:35:13.953811 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:35:13 crc kubenswrapper[4945]: E1206 07:35:13.954884 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:35:27 crc kubenswrapper[4945]: I1206 07:35:27.953726 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:35:27 crc kubenswrapper[4945]: E1206 07:35:27.955122 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:35:40 crc kubenswrapper[4945]: I1206 07:35:40.953467 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:35:40 crc kubenswrapper[4945]: E1206 07:35:40.954587 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:35:51 crc kubenswrapper[4945]: I1206 07:35:51.954087 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:35:51 crc kubenswrapper[4945]: E1206 07:35:51.955439 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:36:06 crc kubenswrapper[4945]: I1206 07:36:06.957431 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:36:06 crc kubenswrapper[4945]: E1206 07:36:06.958437 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:36:18 crc kubenswrapper[4945]: I1206 07:36:18.954203 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:36:18 crc kubenswrapper[4945]: E1206 07:36:18.955650 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:36:32 crc kubenswrapper[4945]: I1206 07:36:32.953342 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:36:32 crc kubenswrapper[4945]: E1206 07:36:32.954430 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:36:45 crc kubenswrapper[4945]: I1206 07:36:45.954130 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:36:45 crc kubenswrapper[4945]: E1206 07:36:45.954824 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:36:58 crc kubenswrapper[4945]: I1206 07:36:58.954005 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:36:58 crc kubenswrapper[4945]: E1206 07:36:58.955397 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:37:12 crc kubenswrapper[4945]: I1206 07:37:12.953917 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:37:12 crc kubenswrapper[4945]: E1206 07:37:12.955113 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:37:23 crc kubenswrapper[4945]: I1206 07:37:23.953531 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:37:23 crc kubenswrapper[4945]: E1206 07:37:23.954633 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:37:37 crc kubenswrapper[4945]: I1206 07:37:37.952999 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:37:37 crc kubenswrapper[4945]: E1206 07:37:37.954790 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:37:52 crc kubenswrapper[4945]: I1206 07:37:52.953818 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:37:52 crc kubenswrapper[4945]: E1206 07:37:52.954800 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:38:05 crc kubenswrapper[4945]: I1206 07:38:05.953132 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:38:05 crc kubenswrapper[4945]: E1206 07:38:05.954414 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:38:16 crc kubenswrapper[4945]: I1206 07:38:16.957315 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:38:17 crc kubenswrapper[4945]: I1206 07:38:17.697607 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"1b2a65ae7ae2276b8dc31bfaaf3b7cee44282759097c6bf0d4ad7864ad1655fd"} Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.608078 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c6vbv"] Dec 06 07:40:20 crc kubenswrapper[4945]: E1206 07:40:20.608909 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerName="registry-server" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.608923 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerName="registry-server" Dec 06 07:40:20 crc kubenswrapper[4945]: E1206 07:40:20.608940 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerName="extract-utilities" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.608947 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerName="extract-utilities" Dec 06 07:40:20 crc kubenswrapper[4945]: E1206 07:40:20.608962 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerName="extract-content" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.608968 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerName="extract-content" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.609109 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef133557-52bb-4cc5-bba4-f44efb51a45d" containerName="registry-server" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.610166 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.623496 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6vbv"] Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.718495 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-catalog-content\") pod \"redhat-marketplace-c6vbv\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.718629 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-utilities\") pod \"redhat-marketplace-c6vbv\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.718687 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mlpw\" (UniqueName: \"kubernetes.io/projected/6257d226-dfd6-48d7-94be-2ee8bf8c832d-kube-api-access-5mlpw\") pod \"redhat-marketplace-c6vbv\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.819573 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-utilities\") pod \"redhat-marketplace-c6vbv\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.819631 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mlpw\" (UniqueName: \"kubernetes.io/projected/6257d226-dfd6-48d7-94be-2ee8bf8c832d-kube-api-access-5mlpw\") pod \"redhat-marketplace-c6vbv\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.819687 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-catalog-content\") pod \"redhat-marketplace-c6vbv\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.820124 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-utilities\") pod \"redhat-marketplace-c6vbv\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.820337 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-catalog-content\") pod \"redhat-marketplace-c6vbv\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.842069 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mlpw\" (UniqueName: \"kubernetes.io/projected/6257d226-dfd6-48d7-94be-2ee8bf8c832d-kube-api-access-5mlpw\") pod \"redhat-marketplace-c6vbv\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:20 crc kubenswrapper[4945]: I1206 07:40:20.936676 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:21 crc kubenswrapper[4945]: I1206 07:40:21.204766 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6vbv"] Dec 06 07:40:21 crc kubenswrapper[4945]: I1206 07:40:21.692067 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6vbv" event={"ID":"6257d226-dfd6-48d7-94be-2ee8bf8c832d","Type":"ContainerStarted","Data":"4707407cae28f0ec2e10758e6ce6448520cabef3220188ef61bdcdabce709a8b"} Dec 06 07:40:22 crc kubenswrapper[4945]: I1206 07:40:22.707554 4945 generic.go:334] "Generic (PLEG): container finished" podID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerID="1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc" exitCode=0 Dec 06 07:40:22 crc kubenswrapper[4945]: I1206 07:40:22.708835 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6vbv" event={"ID":"6257d226-dfd6-48d7-94be-2ee8bf8c832d","Type":"ContainerDied","Data":"1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc"} Dec 06 07:40:22 crc kubenswrapper[4945]: I1206 07:40:22.712883 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:40:24 crc kubenswrapper[4945]: I1206 07:40:24.725942 4945 generic.go:334] "Generic (PLEG): container finished" podID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerID="4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77" exitCode=0 Dec 06 07:40:24 crc kubenswrapper[4945]: I1206 07:40:24.726117 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6vbv" event={"ID":"6257d226-dfd6-48d7-94be-2ee8bf8c832d","Type":"ContainerDied","Data":"4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77"} Dec 06 07:40:26 crc kubenswrapper[4945]: I1206 07:40:26.740142 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6vbv" event={"ID":"6257d226-dfd6-48d7-94be-2ee8bf8c832d","Type":"ContainerStarted","Data":"872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff"} Dec 06 07:40:26 crc kubenswrapper[4945]: I1206 07:40:26.759314 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c6vbv" podStartSLOduration=3.792865569 podStartE2EDuration="6.759273578s" podCreationTimestamp="2025-12-06 07:40:20 +0000 UTC" firstStartedPulling="2025-12-06 07:40:22.712626889 +0000 UTC m=+2836.167487933" lastFinishedPulling="2025-12-06 07:40:25.679034908 +0000 UTC m=+2839.133895942" observedRunningTime="2025-12-06 07:40:26.757397948 +0000 UTC m=+2840.212258992" watchObservedRunningTime="2025-12-06 07:40:26.759273578 +0000 UTC m=+2840.214134612" Dec 06 07:40:30 crc kubenswrapper[4945]: I1206 07:40:30.936964 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:30 crc kubenswrapper[4945]: I1206 07:40:30.938266 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:30 crc kubenswrapper[4945]: I1206 07:40:30.979405 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:31 crc kubenswrapper[4945]: I1206 07:40:31.823694 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:31 crc kubenswrapper[4945]: I1206 07:40:31.867359 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6vbv"] Dec 06 07:40:33 crc kubenswrapper[4945]: I1206 07:40:33.781306 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c6vbv" podUID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerName="registry-server" containerID="cri-o://872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff" gracePeriod=2 Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.710201 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.790754 4945 generic.go:334] "Generic (PLEG): container finished" podID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerID="872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff" exitCode=0 Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.790803 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6vbv" event={"ID":"6257d226-dfd6-48d7-94be-2ee8bf8c832d","Type":"ContainerDied","Data":"872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff"} Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.790835 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6vbv" event={"ID":"6257d226-dfd6-48d7-94be-2ee8bf8c832d","Type":"ContainerDied","Data":"4707407cae28f0ec2e10758e6ce6448520cabef3220188ef61bdcdabce709a8b"} Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.790855 4945 scope.go:117] "RemoveContainer" containerID="872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.790809 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6vbv" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.808317 4945 scope.go:117] "RemoveContainer" containerID="4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.821145 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-utilities\") pod \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.821195 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mlpw\" (UniqueName: \"kubernetes.io/projected/6257d226-dfd6-48d7-94be-2ee8bf8c832d-kube-api-access-5mlpw\") pod \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.821249 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-catalog-content\") pod \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\" (UID: \"6257d226-dfd6-48d7-94be-2ee8bf8c832d\") " Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.823263 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-utilities" (OuterVolumeSpecName: "utilities") pod "6257d226-dfd6-48d7-94be-2ee8bf8c832d" (UID: "6257d226-dfd6-48d7-94be-2ee8bf8c832d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.825053 4945 scope.go:117] "RemoveContainer" containerID="1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.827622 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6257d226-dfd6-48d7-94be-2ee8bf8c832d-kube-api-access-5mlpw" (OuterVolumeSpecName: "kube-api-access-5mlpw") pod "6257d226-dfd6-48d7-94be-2ee8bf8c832d" (UID: "6257d226-dfd6-48d7-94be-2ee8bf8c832d"). InnerVolumeSpecName "kube-api-access-5mlpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.842675 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6257d226-dfd6-48d7-94be-2ee8bf8c832d" (UID: "6257d226-dfd6-48d7-94be-2ee8bf8c832d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.880837 4945 scope.go:117] "RemoveContainer" containerID="872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff" Dec 06 07:40:34 crc kubenswrapper[4945]: E1206 07:40:34.881364 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff\": container with ID starting with 872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff not found: ID does not exist" containerID="872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.881402 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff"} err="failed to get container status \"872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff\": rpc error: code = NotFound desc = could not find container \"872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff\": container with ID starting with 872344d388ecedc957f38e95f9761d98d9af19af53af3fa32fe4ae3139533bff not found: ID does not exist" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.881429 4945 scope.go:117] "RemoveContainer" containerID="4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77" Dec 06 07:40:34 crc kubenswrapper[4945]: E1206 07:40:34.881707 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77\": container with ID starting with 4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77 not found: ID does not exist" containerID="4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.881737 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77"} err="failed to get container status \"4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77\": rpc error: code = NotFound desc = could not find container \"4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77\": container with ID starting with 4a9f3314a3132b00506f581fcad24fa88cf025057397fa52ad33f04783a2eb77 not found: ID does not exist" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.881753 4945 scope.go:117] "RemoveContainer" containerID="1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc" Dec 06 07:40:34 crc kubenswrapper[4945]: E1206 07:40:34.882003 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc\": container with ID starting with 1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc not found: ID does not exist" containerID="1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.882039 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc"} err="failed to get container status \"1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc\": rpc error: code = NotFound desc = could not find container \"1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc\": container with ID starting with 1031e0f0648f245a1fb84bd75ec10fe3af515b209c02c1a1e4fa4648321d7ebc not found: ID does not exist" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.922699 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.922732 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mlpw\" (UniqueName: \"kubernetes.io/projected/6257d226-dfd6-48d7-94be-2ee8bf8c832d-kube-api-access-5mlpw\") on node \"crc\" DevicePath \"\"" Dec 06 07:40:34 crc kubenswrapper[4945]: I1206 07:40:34.922745 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6257d226-dfd6-48d7-94be-2ee8bf8c832d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:40:35 crc kubenswrapper[4945]: I1206 07:40:35.113493 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6vbv"] Dec 06 07:40:35 crc kubenswrapper[4945]: I1206 07:40:35.118682 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6vbv"] Dec 06 07:40:36 crc kubenswrapper[4945]: I1206 07:40:36.962821 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" path="/var/lib/kubelet/pods/6257d226-dfd6-48d7-94be-2ee8bf8c832d/volumes" Dec 06 07:40:38 crc kubenswrapper[4945]: I1206 07:40:38.795961 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:40:38 crc kubenswrapper[4945]: I1206 07:40:38.796327 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.695276 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pnw74"] Dec 06 07:40:49 crc kubenswrapper[4945]: E1206 07:40:49.696723 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerName="extract-content" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.696822 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerName="extract-content" Dec 06 07:40:49 crc kubenswrapper[4945]: E1206 07:40:49.696875 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerName="registry-server" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.696895 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerName="registry-server" Dec 06 07:40:49 crc kubenswrapper[4945]: E1206 07:40:49.696918 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerName="extract-utilities" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.696937 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerName="extract-utilities" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.697397 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6257d226-dfd6-48d7-94be-2ee8bf8c832d" containerName="registry-server" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.700089 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.704802 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pnw74"] Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.735487 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-utilities\") pod \"community-operators-pnw74\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.735592 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmggx\" (UniqueName: \"kubernetes.io/projected/151f3627-5158-4f96-8478-1bacb2ae4e73-kube-api-access-kmggx\") pod \"community-operators-pnw74\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.736025 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-catalog-content\") pod \"community-operators-pnw74\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.837463 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmggx\" (UniqueName: \"kubernetes.io/projected/151f3627-5158-4f96-8478-1bacb2ae4e73-kube-api-access-kmggx\") pod \"community-operators-pnw74\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.837583 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-catalog-content\") pod \"community-operators-pnw74\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.837685 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-utilities\") pod \"community-operators-pnw74\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.838143 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-catalog-content\") pod \"community-operators-pnw74\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.838192 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-utilities\") pod \"community-operators-pnw74\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:49 crc kubenswrapper[4945]: I1206 07:40:49.857875 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmggx\" (UniqueName: \"kubernetes.io/projected/151f3627-5158-4f96-8478-1bacb2ae4e73-kube-api-access-kmggx\") pod \"community-operators-pnw74\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:50 crc kubenswrapper[4945]: I1206 07:40:50.033714 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:40:50 crc kubenswrapper[4945]: I1206 07:40:50.324925 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pnw74"] Dec 06 07:40:50 crc kubenswrapper[4945]: I1206 07:40:50.908188 4945 generic.go:334] "Generic (PLEG): container finished" podID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerID="8383bbd143acb95c0eaab1246f3d6563c8ce62a4e87ce3d36883a5a493004ce0" exitCode=0 Dec 06 07:40:50 crc kubenswrapper[4945]: I1206 07:40:50.908249 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnw74" event={"ID":"151f3627-5158-4f96-8478-1bacb2ae4e73","Type":"ContainerDied","Data":"8383bbd143acb95c0eaab1246f3d6563c8ce62a4e87ce3d36883a5a493004ce0"} Dec 06 07:40:50 crc kubenswrapper[4945]: I1206 07:40:50.908564 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnw74" event={"ID":"151f3627-5158-4f96-8478-1bacb2ae4e73","Type":"ContainerStarted","Data":"51be8bafd413449f4c54f0082378d3d4e6a41514ea735819ee90658d9e36a3a1"} Dec 06 07:40:53 crc kubenswrapper[4945]: I1206 07:40:53.935384 4945 generic.go:334] "Generic (PLEG): container finished" podID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerID="ac3b85e6a4d7370c90dfa35cf3a3e961c0adcdefe820d915ae990f4b30d5ca7b" exitCode=0 Dec 06 07:40:53 crc kubenswrapper[4945]: I1206 07:40:53.935548 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnw74" event={"ID":"151f3627-5158-4f96-8478-1bacb2ae4e73","Type":"ContainerDied","Data":"ac3b85e6a4d7370c90dfa35cf3a3e961c0adcdefe820d915ae990f4b30d5ca7b"} Dec 06 07:40:55 crc kubenswrapper[4945]: I1206 07:40:55.952116 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnw74" event={"ID":"151f3627-5158-4f96-8478-1bacb2ae4e73","Type":"ContainerStarted","Data":"a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5"} Dec 06 07:40:55 crc kubenswrapper[4945]: I1206 07:40:55.972444 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pnw74" podStartSLOduration=3.07809178 podStartE2EDuration="6.972419271s" podCreationTimestamp="2025-12-06 07:40:49 +0000 UTC" firstStartedPulling="2025-12-06 07:40:50.910085159 +0000 UTC m=+2864.364946213" lastFinishedPulling="2025-12-06 07:40:54.80441266 +0000 UTC m=+2868.259273704" observedRunningTime="2025-12-06 07:40:55.970222242 +0000 UTC m=+2869.425083296" watchObservedRunningTime="2025-12-06 07:40:55.972419271 +0000 UTC m=+2869.427280335" Dec 06 07:41:00 crc kubenswrapper[4945]: I1206 07:41:00.034496 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:41:00 crc kubenswrapper[4945]: I1206 07:41:00.034920 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:41:00 crc kubenswrapper[4945]: I1206 07:41:00.111595 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:41:01 crc kubenswrapper[4945]: I1206 07:41:01.068525 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:41:03 crc kubenswrapper[4945]: I1206 07:41:03.880325 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pnw74"] Dec 06 07:41:03 crc kubenswrapper[4945]: I1206 07:41:03.880936 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pnw74" podUID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerName="registry-server" containerID="cri-o://a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5" gracePeriod=2 Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:08.796648 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:08.797463 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:41:15 crc kubenswrapper[4945]: E1206 07:41:10.034934 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5 is running failed: container process not found" containerID="a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 07:41:15 crc kubenswrapper[4945]: E1206 07:41:10.035464 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5 is running failed: container process not found" containerID="a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 07:41:15 crc kubenswrapper[4945]: E1206 07:41:10.036038 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5 is running failed: container process not found" containerID="a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 07:41:15 crc kubenswrapper[4945]: E1206 07:41:10.036093 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-pnw74" podUID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerName="registry-server" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:11.807189 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pnw74_151f3627-5158-4f96-8478-1bacb2ae4e73/registry-server/0.log" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:11.808744 4945 generic.go:334] "Generic (PLEG): container finished" podID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerID="a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5" exitCode=-1 Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:11.808828 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnw74" event={"ID":"151f3627-5158-4f96-8478-1bacb2ae4e73","Type":"ContainerDied","Data":"a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5"} Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.331192 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.447190 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmggx\" (UniqueName: \"kubernetes.io/projected/151f3627-5158-4f96-8478-1bacb2ae4e73-kube-api-access-kmggx\") pod \"151f3627-5158-4f96-8478-1bacb2ae4e73\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.447246 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-catalog-content\") pod \"151f3627-5158-4f96-8478-1bacb2ae4e73\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.447408 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-utilities\") pod \"151f3627-5158-4f96-8478-1bacb2ae4e73\" (UID: \"151f3627-5158-4f96-8478-1bacb2ae4e73\") " Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.449104 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-utilities" (OuterVolumeSpecName: "utilities") pod "151f3627-5158-4f96-8478-1bacb2ae4e73" (UID: "151f3627-5158-4f96-8478-1bacb2ae4e73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.453151 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/151f3627-5158-4f96-8478-1bacb2ae4e73-kube-api-access-kmggx" (OuterVolumeSpecName: "kube-api-access-kmggx") pod "151f3627-5158-4f96-8478-1bacb2ae4e73" (UID: "151f3627-5158-4f96-8478-1bacb2ae4e73"). InnerVolumeSpecName "kube-api-access-kmggx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.505411 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "151f3627-5158-4f96-8478-1bacb2ae4e73" (UID: "151f3627-5158-4f96-8478-1bacb2ae4e73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.550271 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmggx\" (UniqueName: \"kubernetes.io/projected/151f3627-5158-4f96-8478-1bacb2ae4e73-kube-api-access-kmggx\") on node \"crc\" DevicePath \"\"" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.550315 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.550325 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/151f3627-5158-4f96-8478-1bacb2ae4e73-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.856195 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnw74" event={"ID":"151f3627-5158-4f96-8478-1bacb2ae4e73","Type":"ContainerDied","Data":"51be8bafd413449f4c54f0082378d3d4e6a41514ea735819ee90658d9e36a3a1"} Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.856246 4945 scope.go:117] "RemoveContainer" containerID="a7a17dfd4e8fe4bcc6502be30b76cc55c8c698585e70ed3cba8e95948de5e5e5" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.856328 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pnw74" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.875814 4945 scope.go:117] "RemoveContainer" containerID="ac3b85e6a4d7370c90dfa35cf3a3e961c0adcdefe820d915ae990f4b30d5ca7b" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.898272 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pnw74"] Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.907430 4945 scope.go:117] "RemoveContainer" containerID="8383bbd143acb95c0eaab1246f3d6563c8ce62a4e87ce3d36883a5a493004ce0" Dec 06 07:41:15 crc kubenswrapper[4945]: I1206 07:41:15.908269 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pnw74"] Dec 06 07:41:16 crc kubenswrapper[4945]: I1206 07:41:16.962360 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="151f3627-5158-4f96-8478-1bacb2ae4e73" path="/var/lib/kubelet/pods/151f3627-5158-4f96-8478-1bacb2ae4e73/volumes" Dec 06 07:41:38 crc kubenswrapper[4945]: I1206 07:41:38.795911 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:41:38 crc kubenswrapper[4945]: I1206 07:41:38.796618 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:41:38 crc kubenswrapper[4945]: I1206 07:41:38.796666 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:41:38 crc kubenswrapper[4945]: I1206 07:41:38.797308 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b2a65ae7ae2276b8dc31bfaaf3b7cee44282759097c6bf0d4ad7864ad1655fd"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:41:38 crc kubenswrapper[4945]: I1206 07:41:38.797375 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://1b2a65ae7ae2276b8dc31bfaaf3b7cee44282759097c6bf0d4ad7864ad1655fd" gracePeriod=600 Dec 06 07:41:40 crc kubenswrapper[4945]: I1206 07:41:40.042364 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="1b2a65ae7ae2276b8dc31bfaaf3b7cee44282759097c6bf0d4ad7864ad1655fd" exitCode=0 Dec 06 07:41:40 crc kubenswrapper[4945]: I1206 07:41:40.042564 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"1b2a65ae7ae2276b8dc31bfaaf3b7cee44282759097c6bf0d4ad7864ad1655fd"} Dec 06 07:41:40 crc kubenswrapper[4945]: I1206 07:41:40.042893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e"} Dec 06 07:41:40 crc kubenswrapper[4945]: I1206 07:41:40.042914 4945 scope.go:117] "RemoveContainer" containerID="b3f662a0db4a7a47a1c0e84c80d42a86ad65baa09e42cf09580558bd883bf000" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.192894 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z65g6"] Dec 06 07:42:22 crc kubenswrapper[4945]: E1206 07:42:22.193749 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerName="extract-utilities" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.193765 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerName="extract-utilities" Dec 06 07:42:22 crc kubenswrapper[4945]: E1206 07:42:22.193792 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerName="extract-content" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.193800 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerName="extract-content" Dec 06 07:42:22 crc kubenswrapper[4945]: E1206 07:42:22.193820 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerName="registry-server" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.193831 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerName="registry-server" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.194008 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="151f3627-5158-4f96-8478-1bacb2ae4e73" containerName="registry-server" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.195427 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.221889 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z65g6"] Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.243046 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-catalog-content\") pod \"redhat-operators-z65g6\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.243383 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-utilities\") pod \"redhat-operators-z65g6\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.243535 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtqkz\" (UniqueName: \"kubernetes.io/projected/a51037b9-0aae-45cc-b257-73d574724ef2-kube-api-access-vtqkz\") pod \"redhat-operators-z65g6\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.344299 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtqkz\" (UniqueName: \"kubernetes.io/projected/a51037b9-0aae-45cc-b257-73d574724ef2-kube-api-access-vtqkz\") pod \"redhat-operators-z65g6\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.344400 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-catalog-content\") pod \"redhat-operators-z65g6\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.344475 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-utilities\") pod \"redhat-operators-z65g6\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.345038 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-catalog-content\") pod \"redhat-operators-z65g6\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.345139 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-utilities\") pod \"redhat-operators-z65g6\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.372403 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtqkz\" (UniqueName: \"kubernetes.io/projected/a51037b9-0aae-45cc-b257-73d574724ef2-kube-api-access-vtqkz\") pod \"redhat-operators-z65g6\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.513114 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:22 crc kubenswrapper[4945]: I1206 07:42:22.961550 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z65g6"] Dec 06 07:42:23 crc kubenswrapper[4945]: I1206 07:42:23.377936 4945 generic.go:334] "Generic (PLEG): container finished" podID="a51037b9-0aae-45cc-b257-73d574724ef2" containerID="dae69c32dce80146cbe171522175357584d61334495763d5c668122ffd27fb9a" exitCode=0 Dec 06 07:42:23 crc kubenswrapper[4945]: I1206 07:42:23.378487 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z65g6" event={"ID":"a51037b9-0aae-45cc-b257-73d574724ef2","Type":"ContainerDied","Data":"dae69c32dce80146cbe171522175357584d61334495763d5c668122ffd27fb9a"} Dec 06 07:42:23 crc kubenswrapper[4945]: I1206 07:42:23.378525 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z65g6" event={"ID":"a51037b9-0aae-45cc-b257-73d574724ef2","Type":"ContainerStarted","Data":"42eedf02317b1233d4d1cfdcb7d57fedf52b271a3adfb4aab32a834d81691a44"} Dec 06 07:42:24 crc kubenswrapper[4945]: I1206 07:42:24.386135 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z65g6" event={"ID":"a51037b9-0aae-45cc-b257-73d574724ef2","Type":"ContainerStarted","Data":"88abb79ddcf7b3edac1c95ffc119b97a8cb4e7d945ed5e4bb2f42bbd4b04978c"} Dec 06 07:42:25 crc kubenswrapper[4945]: I1206 07:42:25.395048 4945 generic.go:334] "Generic (PLEG): container finished" podID="a51037b9-0aae-45cc-b257-73d574724ef2" containerID="88abb79ddcf7b3edac1c95ffc119b97a8cb4e7d945ed5e4bb2f42bbd4b04978c" exitCode=0 Dec 06 07:42:25 crc kubenswrapper[4945]: I1206 07:42:25.395179 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z65g6" event={"ID":"a51037b9-0aae-45cc-b257-73d574724ef2","Type":"ContainerDied","Data":"88abb79ddcf7b3edac1c95ffc119b97a8cb4e7d945ed5e4bb2f42bbd4b04978c"} Dec 06 07:42:26 crc kubenswrapper[4945]: I1206 07:42:26.403598 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z65g6" event={"ID":"a51037b9-0aae-45cc-b257-73d574724ef2","Type":"ContainerStarted","Data":"6ae59d7abbdf6438d073450b57b48c216c1d50216f5e770ea2bc4829523b093d"} Dec 06 07:42:26 crc kubenswrapper[4945]: I1206 07:42:26.428647 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z65g6" podStartSLOduration=2.018576926 podStartE2EDuration="4.428625841s" podCreationTimestamp="2025-12-06 07:42:22 +0000 UTC" firstStartedPulling="2025-12-06 07:42:23.380055539 +0000 UTC m=+2956.834916583" lastFinishedPulling="2025-12-06 07:42:25.790104444 +0000 UTC m=+2959.244965498" observedRunningTime="2025-12-06 07:42:26.420645507 +0000 UTC m=+2959.875506571" watchObservedRunningTime="2025-12-06 07:42:26.428625841 +0000 UTC m=+2959.883486885" Dec 06 07:42:32 crc kubenswrapper[4945]: I1206 07:42:32.514245 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:32 crc kubenswrapper[4945]: I1206 07:42:32.514953 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:32 crc kubenswrapper[4945]: I1206 07:42:32.604923 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:33 crc kubenswrapper[4945]: I1206 07:42:33.543944 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:33 crc kubenswrapper[4945]: I1206 07:42:33.625347 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z65g6"] Dec 06 07:42:35 crc kubenswrapper[4945]: I1206 07:42:35.475231 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z65g6" podUID="a51037b9-0aae-45cc-b257-73d574724ef2" containerName="registry-server" containerID="cri-o://6ae59d7abbdf6438d073450b57b48c216c1d50216f5e770ea2bc4829523b093d" gracePeriod=2 Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.494823 4945 generic.go:334] "Generic (PLEG): container finished" podID="a51037b9-0aae-45cc-b257-73d574724ef2" containerID="6ae59d7abbdf6438d073450b57b48c216c1d50216f5e770ea2bc4829523b093d" exitCode=0 Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.494925 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z65g6" event={"ID":"a51037b9-0aae-45cc-b257-73d574724ef2","Type":"ContainerDied","Data":"6ae59d7abbdf6438d073450b57b48c216c1d50216f5e770ea2bc4829523b093d"} Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.724344 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.872539 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-catalog-content\") pod \"a51037b9-0aae-45cc-b257-73d574724ef2\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.872632 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-utilities\") pod \"a51037b9-0aae-45cc-b257-73d574724ef2\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.872742 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtqkz\" (UniqueName: \"kubernetes.io/projected/a51037b9-0aae-45cc-b257-73d574724ef2-kube-api-access-vtqkz\") pod \"a51037b9-0aae-45cc-b257-73d574724ef2\" (UID: \"a51037b9-0aae-45cc-b257-73d574724ef2\") " Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.874230 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-utilities" (OuterVolumeSpecName: "utilities") pod "a51037b9-0aae-45cc-b257-73d574724ef2" (UID: "a51037b9-0aae-45cc-b257-73d574724ef2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.880081 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a51037b9-0aae-45cc-b257-73d574724ef2-kube-api-access-vtqkz" (OuterVolumeSpecName: "kube-api-access-vtqkz") pod "a51037b9-0aae-45cc-b257-73d574724ef2" (UID: "a51037b9-0aae-45cc-b257-73d574724ef2"). InnerVolumeSpecName "kube-api-access-vtqkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.974591 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtqkz\" (UniqueName: \"kubernetes.io/projected/a51037b9-0aae-45cc-b257-73d574724ef2-kube-api-access-vtqkz\") on node \"crc\" DevicePath \"\"" Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.974620 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:42:37 crc kubenswrapper[4945]: I1206 07:42:37.987861 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a51037b9-0aae-45cc-b257-73d574724ef2" (UID: "a51037b9-0aae-45cc-b257-73d574724ef2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:42:38 crc kubenswrapper[4945]: I1206 07:42:38.076424 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a51037b9-0aae-45cc-b257-73d574724ef2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:42:38 crc kubenswrapper[4945]: I1206 07:42:38.505789 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z65g6" event={"ID":"a51037b9-0aae-45cc-b257-73d574724ef2","Type":"ContainerDied","Data":"42eedf02317b1233d4d1cfdcb7d57fedf52b271a3adfb4aab32a834d81691a44"} Dec 06 07:42:38 crc kubenswrapper[4945]: I1206 07:42:38.505841 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z65g6" Dec 06 07:42:38 crc kubenswrapper[4945]: I1206 07:42:38.505871 4945 scope.go:117] "RemoveContainer" containerID="6ae59d7abbdf6438d073450b57b48c216c1d50216f5e770ea2bc4829523b093d" Dec 06 07:42:38 crc kubenswrapper[4945]: I1206 07:42:38.527054 4945 scope.go:117] "RemoveContainer" containerID="88abb79ddcf7b3edac1c95ffc119b97a8cb4e7d945ed5e4bb2f42bbd4b04978c" Dec 06 07:42:38 crc kubenswrapper[4945]: I1206 07:42:38.542228 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z65g6"] Dec 06 07:42:38 crc kubenswrapper[4945]: I1206 07:42:38.553482 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z65g6"] Dec 06 07:42:38 crc kubenswrapper[4945]: I1206 07:42:38.567100 4945 scope.go:117] "RemoveContainer" containerID="dae69c32dce80146cbe171522175357584d61334495763d5c668122ffd27fb9a" Dec 06 07:42:38 crc kubenswrapper[4945]: I1206 07:42:38.970000 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a51037b9-0aae-45cc-b257-73d574724ef2" path="/var/lib/kubelet/pods/a51037b9-0aae-45cc-b257-73d574724ef2/volumes" Dec 06 07:44:08 crc kubenswrapper[4945]: I1206 07:44:08.795852 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:44:08 crc kubenswrapper[4945]: I1206 07:44:08.797243 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:44:38 crc kubenswrapper[4945]: I1206 07:44:38.796412 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:44:38 crc kubenswrapper[4945]: I1206 07:44:38.797718 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.178109 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz"] Dec 06 07:45:00 crc kubenswrapper[4945]: E1206 07:45:00.179083 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51037b9-0aae-45cc-b257-73d574724ef2" containerName="registry-server" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.179101 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51037b9-0aae-45cc-b257-73d574724ef2" containerName="registry-server" Dec 06 07:45:00 crc kubenswrapper[4945]: E1206 07:45:00.179133 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51037b9-0aae-45cc-b257-73d574724ef2" containerName="extract-utilities" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.179141 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51037b9-0aae-45cc-b257-73d574724ef2" containerName="extract-utilities" Dec 06 07:45:00 crc kubenswrapper[4945]: E1206 07:45:00.179150 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51037b9-0aae-45cc-b257-73d574724ef2" containerName="extract-content" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.179157 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51037b9-0aae-45cc-b257-73d574724ef2" containerName="extract-content" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.179434 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a51037b9-0aae-45cc-b257-73d574724ef2" containerName="registry-server" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.180046 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.184873 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz"] Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.189042 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.189237 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.260631 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4bbx\" (UniqueName: \"kubernetes.io/projected/44485cbb-739b-4d1b-90bd-7a413cff1896-kube-api-access-r4bbx\") pod \"collect-profiles-29416785-znwrz\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.260676 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44485cbb-739b-4d1b-90bd-7a413cff1896-secret-volume\") pod \"collect-profiles-29416785-znwrz\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.260720 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44485cbb-739b-4d1b-90bd-7a413cff1896-config-volume\") pod \"collect-profiles-29416785-znwrz\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.362176 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44485cbb-739b-4d1b-90bd-7a413cff1896-secret-volume\") pod \"collect-profiles-29416785-znwrz\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.362244 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44485cbb-739b-4d1b-90bd-7a413cff1896-config-volume\") pod \"collect-profiles-29416785-znwrz\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.362331 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4bbx\" (UniqueName: \"kubernetes.io/projected/44485cbb-739b-4d1b-90bd-7a413cff1896-kube-api-access-r4bbx\") pod \"collect-profiles-29416785-znwrz\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.363590 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44485cbb-739b-4d1b-90bd-7a413cff1896-config-volume\") pod \"collect-profiles-29416785-znwrz\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.373628 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44485cbb-739b-4d1b-90bd-7a413cff1896-secret-volume\") pod \"collect-profiles-29416785-znwrz\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.377582 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4bbx\" (UniqueName: \"kubernetes.io/projected/44485cbb-739b-4d1b-90bd-7a413cff1896-kube-api-access-r4bbx\") pod \"collect-profiles-29416785-znwrz\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.504153 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:00 crc kubenswrapper[4945]: I1206 07:45:00.969342 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz"] Dec 06 07:45:01 crc kubenswrapper[4945]: I1206 07:45:01.747261 4945 generic.go:334] "Generic (PLEG): container finished" podID="44485cbb-739b-4d1b-90bd-7a413cff1896" containerID="ac7364973e8403f04634ad22bb330c62c86967a21c8c84bef280f0792f020922" exitCode=0 Dec 06 07:45:01 crc kubenswrapper[4945]: I1206 07:45:01.747405 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" event={"ID":"44485cbb-739b-4d1b-90bd-7a413cff1896","Type":"ContainerDied","Data":"ac7364973e8403f04634ad22bb330c62c86967a21c8c84bef280f0792f020922"} Dec 06 07:45:01 crc kubenswrapper[4945]: I1206 07:45:01.748525 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" event={"ID":"44485cbb-739b-4d1b-90bd-7a413cff1896","Type":"ContainerStarted","Data":"e5280fa368763e2b87b70694710ba6e7664d132c6c32b4f297b45530ddec9c76"} Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.009918 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.098477 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44485cbb-739b-4d1b-90bd-7a413cff1896-secret-volume\") pod \"44485cbb-739b-4d1b-90bd-7a413cff1896\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.098528 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44485cbb-739b-4d1b-90bd-7a413cff1896-config-volume\") pod \"44485cbb-739b-4d1b-90bd-7a413cff1896\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.098623 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4bbx\" (UniqueName: \"kubernetes.io/projected/44485cbb-739b-4d1b-90bd-7a413cff1896-kube-api-access-r4bbx\") pod \"44485cbb-739b-4d1b-90bd-7a413cff1896\" (UID: \"44485cbb-739b-4d1b-90bd-7a413cff1896\") " Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.099275 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44485cbb-739b-4d1b-90bd-7a413cff1896-config-volume" (OuterVolumeSpecName: "config-volume") pod "44485cbb-739b-4d1b-90bd-7a413cff1896" (UID: "44485cbb-739b-4d1b-90bd-7a413cff1896"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.104410 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44485cbb-739b-4d1b-90bd-7a413cff1896-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "44485cbb-739b-4d1b-90bd-7a413cff1896" (UID: "44485cbb-739b-4d1b-90bd-7a413cff1896"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.104492 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44485cbb-739b-4d1b-90bd-7a413cff1896-kube-api-access-r4bbx" (OuterVolumeSpecName: "kube-api-access-r4bbx") pod "44485cbb-739b-4d1b-90bd-7a413cff1896" (UID: "44485cbb-739b-4d1b-90bd-7a413cff1896"). InnerVolumeSpecName "kube-api-access-r4bbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.200611 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4bbx\" (UniqueName: \"kubernetes.io/projected/44485cbb-739b-4d1b-90bd-7a413cff1896-kube-api-access-r4bbx\") on node \"crc\" DevicePath \"\"" Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.200674 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44485cbb-739b-4d1b-90bd-7a413cff1896-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.200688 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44485cbb-739b-4d1b-90bd-7a413cff1896-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.763820 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" event={"ID":"44485cbb-739b-4d1b-90bd-7a413cff1896","Type":"ContainerDied","Data":"e5280fa368763e2b87b70694710ba6e7664d132c6c32b4f297b45530ddec9c76"} Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.764351 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5280fa368763e2b87b70694710ba6e7664d132c6c32b4f297b45530ddec9c76" Dec 06 07:45:03 crc kubenswrapper[4945]: I1206 07:45:03.764202 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz" Dec 06 07:45:04 crc kubenswrapper[4945]: I1206 07:45:04.076359 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw"] Dec 06 07:45:04 crc kubenswrapper[4945]: I1206 07:45:04.097112 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-mfhxw"] Dec 06 07:45:04 crc kubenswrapper[4945]: I1206 07:45:04.963030 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c7767b5-806a-47a0-8ec6-be3e9fdeca18" path="/var/lib/kubelet/pods/7c7767b5-806a-47a0-8ec6-be3e9fdeca18/volumes" Dec 06 07:45:08 crc kubenswrapper[4945]: I1206 07:45:08.795823 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:45:08 crc kubenswrapper[4945]: I1206 07:45:08.796186 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:45:08 crc kubenswrapper[4945]: I1206 07:45:08.796231 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:45:08 crc kubenswrapper[4945]: I1206 07:45:08.796864 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:45:08 crc kubenswrapper[4945]: I1206 07:45:08.796920 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" gracePeriod=600 Dec 06 07:45:08 crc kubenswrapper[4945]: E1206 07:45:08.925711 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:45:09 crc kubenswrapper[4945]: I1206 07:45:09.813644 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" exitCode=0 Dec 06 07:45:09 crc kubenswrapper[4945]: I1206 07:45:09.813696 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e"} Dec 06 07:45:09 crc kubenswrapper[4945]: I1206 07:45:09.813733 4945 scope.go:117] "RemoveContainer" containerID="1b2a65ae7ae2276b8dc31bfaaf3b7cee44282759097c6bf0d4ad7864ad1655fd" Dec 06 07:45:09 crc kubenswrapper[4945]: I1206 07:45:09.814241 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:45:09 crc kubenswrapper[4945]: E1206 07:45:09.814466 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:45:18 crc kubenswrapper[4945]: I1206 07:45:18.030177 4945 scope.go:117] "RemoveContainer" containerID="e13824d49ddbecada56e31c77cd25010844935e9d016a2d8e0ba5c61c06f381d" Dec 06 07:45:22 crc kubenswrapper[4945]: I1206 07:45:22.953489 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:45:22 crc kubenswrapper[4945]: E1206 07:45:22.954191 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:45:36 crc kubenswrapper[4945]: I1206 07:45:36.958262 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:45:36 crc kubenswrapper[4945]: E1206 07:45:36.958997 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:45:50 crc kubenswrapper[4945]: I1206 07:45:50.953576 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:45:50 crc kubenswrapper[4945]: E1206 07:45:50.954468 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:46:05 crc kubenswrapper[4945]: I1206 07:46:05.953188 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:46:05 crc kubenswrapper[4945]: E1206 07:46:05.954066 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:46:17 crc kubenswrapper[4945]: I1206 07:46:17.953057 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:46:17 crc kubenswrapper[4945]: E1206 07:46:17.954334 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:46:30 crc kubenswrapper[4945]: I1206 07:46:30.953345 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:46:30 crc kubenswrapper[4945]: E1206 07:46:30.954251 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:46:45 crc kubenswrapper[4945]: I1206 07:46:45.953728 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:46:45 crc kubenswrapper[4945]: E1206 07:46:45.954612 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:46:58 crc kubenswrapper[4945]: I1206 07:46:58.953017 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:46:58 crc kubenswrapper[4945]: E1206 07:46:58.954000 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:47:11 crc kubenswrapper[4945]: I1206 07:47:11.952689 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:47:11 crc kubenswrapper[4945]: E1206 07:47:11.953564 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:47:24 crc kubenswrapper[4945]: I1206 07:47:24.953524 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:47:24 crc kubenswrapper[4945]: E1206 07:47:24.954555 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:47:39 crc kubenswrapper[4945]: I1206 07:47:39.953611 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:47:39 crc kubenswrapper[4945]: E1206 07:47:39.954725 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:47:53 crc kubenswrapper[4945]: I1206 07:47:53.952891 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:47:53 crc kubenswrapper[4945]: E1206 07:47:53.953520 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.262305 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pnnfx"] Dec 06 07:47:59 crc kubenswrapper[4945]: E1206 07:47:59.263430 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44485cbb-739b-4d1b-90bd-7a413cff1896" containerName="collect-profiles" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.263454 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="44485cbb-739b-4d1b-90bd-7a413cff1896" containerName="collect-profiles" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.263730 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="44485cbb-739b-4d1b-90bd-7a413cff1896" containerName="collect-profiles" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.265053 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.269888 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pnnfx"] Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.458080 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-catalog-content\") pod \"certified-operators-pnnfx\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.458171 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-utilities\") pod \"certified-operators-pnnfx\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.458208 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmdxr\" (UniqueName: \"kubernetes.io/projected/20eed229-33e8-4bf8-9620-c797dfab6350-kube-api-access-gmdxr\") pod \"certified-operators-pnnfx\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.559314 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-catalog-content\") pod \"certified-operators-pnnfx\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.559363 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-utilities\") pod \"certified-operators-pnnfx\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.559391 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmdxr\" (UniqueName: \"kubernetes.io/projected/20eed229-33e8-4bf8-9620-c797dfab6350-kube-api-access-gmdxr\") pod \"certified-operators-pnnfx\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.560402 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-utilities\") pod \"certified-operators-pnnfx\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.560438 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-catalog-content\") pod \"certified-operators-pnnfx\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.583380 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmdxr\" (UniqueName: \"kubernetes.io/projected/20eed229-33e8-4bf8-9620-c797dfab6350-kube-api-access-gmdxr\") pod \"certified-operators-pnnfx\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.589519 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:47:59 crc kubenswrapper[4945]: I1206 07:47:59.841642 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pnnfx"] Dec 06 07:48:00 crc kubenswrapper[4945]: I1206 07:48:00.173565 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnnfx" event={"ID":"20eed229-33e8-4bf8-9620-c797dfab6350","Type":"ContainerStarted","Data":"693ff85b2e1396f9eff90ffc8fb0c763cc269a043937f687ce7861d9ef13e0b2"} Dec 06 07:48:02 crc kubenswrapper[4945]: I1206 07:48:02.190925 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnnfx" event={"ID":"20eed229-33e8-4bf8-9620-c797dfab6350","Type":"ContainerStarted","Data":"1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573"} Dec 06 07:48:03 crc kubenswrapper[4945]: I1206 07:48:03.199398 4945 generic.go:334] "Generic (PLEG): container finished" podID="20eed229-33e8-4bf8-9620-c797dfab6350" containerID="1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573" exitCode=0 Dec 06 07:48:03 crc kubenswrapper[4945]: I1206 07:48:03.199437 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnnfx" event={"ID":"20eed229-33e8-4bf8-9620-c797dfab6350","Type":"ContainerDied","Data":"1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573"} Dec 06 07:48:03 crc kubenswrapper[4945]: I1206 07:48:03.202045 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:48:05 crc kubenswrapper[4945]: I1206 07:48:05.953873 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:48:05 crc kubenswrapper[4945]: E1206 07:48:05.954467 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:48:06 crc kubenswrapper[4945]: I1206 07:48:06.224235 4945 generic.go:334] "Generic (PLEG): container finished" podID="20eed229-33e8-4bf8-9620-c797dfab6350" containerID="129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a" exitCode=0 Dec 06 07:48:06 crc kubenswrapper[4945]: I1206 07:48:06.224292 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnnfx" event={"ID":"20eed229-33e8-4bf8-9620-c797dfab6350","Type":"ContainerDied","Data":"129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a"} Dec 06 07:48:07 crc kubenswrapper[4945]: I1206 07:48:07.232882 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnnfx" event={"ID":"20eed229-33e8-4bf8-9620-c797dfab6350","Type":"ContainerStarted","Data":"fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4"} Dec 06 07:48:07 crc kubenswrapper[4945]: I1206 07:48:07.255408 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pnnfx" podStartSLOduration=4.709296782 podStartE2EDuration="8.255389963s" podCreationTimestamp="2025-12-06 07:47:59 +0000 UTC" firstStartedPulling="2025-12-06 07:48:03.201534741 +0000 UTC m=+3296.656395785" lastFinishedPulling="2025-12-06 07:48:06.747627922 +0000 UTC m=+3300.202488966" observedRunningTime="2025-12-06 07:48:07.250361929 +0000 UTC m=+3300.705222993" watchObservedRunningTime="2025-12-06 07:48:07.255389963 +0000 UTC m=+3300.710251007" Dec 06 07:48:09 crc kubenswrapper[4945]: I1206 07:48:09.590335 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:48:09 crc kubenswrapper[4945]: I1206 07:48:09.590629 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:48:09 crc kubenswrapper[4945]: I1206 07:48:09.631889 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:48:16 crc kubenswrapper[4945]: I1206 07:48:16.958602 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:48:16 crc kubenswrapper[4945]: E1206 07:48:16.959441 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:48:19 crc kubenswrapper[4945]: I1206 07:48:19.628396 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:48:19 crc kubenswrapper[4945]: I1206 07:48:19.675253 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pnnfx"] Dec 06 07:48:20 crc kubenswrapper[4945]: I1206 07:48:20.337832 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pnnfx" podUID="20eed229-33e8-4bf8-9620-c797dfab6350" containerName="registry-server" containerID="cri-o://fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4" gracePeriod=2 Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.252245 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.346632 4945 generic.go:334] "Generic (PLEG): container finished" podID="20eed229-33e8-4bf8-9620-c797dfab6350" containerID="fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4" exitCode=0 Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.346671 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnnfx" event={"ID":"20eed229-33e8-4bf8-9620-c797dfab6350","Type":"ContainerDied","Data":"fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4"} Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.346682 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnnfx" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.346702 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnnfx" event={"ID":"20eed229-33e8-4bf8-9620-c797dfab6350","Type":"ContainerDied","Data":"693ff85b2e1396f9eff90ffc8fb0c763cc269a043937f687ce7861d9ef13e0b2"} Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.346725 4945 scope.go:117] "RemoveContainer" containerID="fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.364072 4945 scope.go:117] "RemoveContainer" containerID="129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.378823 4945 scope.go:117] "RemoveContainer" containerID="1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.402479 4945 scope.go:117] "RemoveContainer" containerID="fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4" Dec 06 07:48:21 crc kubenswrapper[4945]: E1206 07:48:21.402947 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4\": container with ID starting with fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4 not found: ID does not exist" containerID="fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.402977 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4"} err="failed to get container status \"fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4\": rpc error: code = NotFound desc = could not find container \"fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4\": container with ID starting with fb2d6666a0d1b991fd3eb21278d3a643b4393bb9ae097c9403623b4558ba5ae4 not found: ID does not exist" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.402997 4945 scope.go:117] "RemoveContainer" containerID="129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a" Dec 06 07:48:21 crc kubenswrapper[4945]: E1206 07:48:21.403261 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a\": container with ID starting with 129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a not found: ID does not exist" containerID="129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.403297 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a"} err="failed to get container status \"129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a\": rpc error: code = NotFound desc = could not find container \"129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a\": container with ID starting with 129a79816b5c63258a74c0059e413e550d6762f870a4a0c13e81f9f472aa096a not found: ID does not exist" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.403316 4945 scope.go:117] "RemoveContainer" containerID="1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573" Dec 06 07:48:21 crc kubenswrapper[4945]: E1206 07:48:21.403673 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573\": container with ID starting with 1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573 not found: ID does not exist" containerID="1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.403699 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573"} err="failed to get container status \"1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573\": rpc error: code = NotFound desc = could not find container \"1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573\": container with ID starting with 1cb90409a2127602d2b934899f3c78a4b672936719750422c4ca344378cb7573 not found: ID does not exist" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.416551 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-utilities\") pod \"20eed229-33e8-4bf8-9620-c797dfab6350\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.416664 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmdxr\" (UniqueName: \"kubernetes.io/projected/20eed229-33e8-4bf8-9620-c797dfab6350-kube-api-access-gmdxr\") pod \"20eed229-33e8-4bf8-9620-c797dfab6350\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.416708 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-catalog-content\") pod \"20eed229-33e8-4bf8-9620-c797dfab6350\" (UID: \"20eed229-33e8-4bf8-9620-c797dfab6350\") " Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.420509 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-utilities" (OuterVolumeSpecName: "utilities") pod "20eed229-33e8-4bf8-9620-c797dfab6350" (UID: "20eed229-33e8-4bf8-9620-c797dfab6350"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.424361 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20eed229-33e8-4bf8-9620-c797dfab6350-kube-api-access-gmdxr" (OuterVolumeSpecName: "kube-api-access-gmdxr") pod "20eed229-33e8-4bf8-9620-c797dfab6350" (UID: "20eed229-33e8-4bf8-9620-c797dfab6350"). InnerVolumeSpecName "kube-api-access-gmdxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.470482 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20eed229-33e8-4bf8-9620-c797dfab6350" (UID: "20eed229-33e8-4bf8-9620-c797dfab6350"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.517926 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.517960 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmdxr\" (UniqueName: \"kubernetes.io/projected/20eed229-33e8-4bf8-9620-c797dfab6350-kube-api-access-gmdxr\") on node \"crc\" DevicePath \"\"" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.517970 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20eed229-33e8-4bf8-9620-c797dfab6350-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.678699 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pnnfx"] Dec 06 07:48:21 crc kubenswrapper[4945]: I1206 07:48:21.684139 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pnnfx"] Dec 06 07:48:22 crc kubenswrapper[4945]: I1206 07:48:22.964532 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20eed229-33e8-4bf8-9620-c797dfab6350" path="/var/lib/kubelet/pods/20eed229-33e8-4bf8-9620-c797dfab6350/volumes" Dec 06 07:48:30 crc kubenswrapper[4945]: I1206 07:48:30.953648 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:48:30 crc kubenswrapper[4945]: E1206 07:48:30.954418 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:48:41 crc kubenswrapper[4945]: I1206 07:48:41.953401 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:48:41 crc kubenswrapper[4945]: E1206 07:48:41.954021 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:48:52 crc kubenswrapper[4945]: I1206 07:48:52.956083 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:48:52 crc kubenswrapper[4945]: E1206 07:48:52.957025 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:49:03 crc kubenswrapper[4945]: I1206 07:49:03.952937 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:49:03 crc kubenswrapper[4945]: E1206 07:49:03.953842 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:49:14 crc kubenswrapper[4945]: I1206 07:49:14.953943 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:49:14 crc kubenswrapper[4945]: E1206 07:49:14.955117 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:49:27 crc kubenswrapper[4945]: I1206 07:49:27.954037 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:49:27 crc kubenswrapper[4945]: E1206 07:49:27.955089 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:49:40 crc kubenswrapper[4945]: I1206 07:49:40.953594 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:49:40 crc kubenswrapper[4945]: E1206 07:49:40.954395 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:49:53 crc kubenswrapper[4945]: I1206 07:49:53.953099 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:49:53 crc kubenswrapper[4945]: E1206 07:49:53.956138 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:50:08 crc kubenswrapper[4945]: I1206 07:50:08.953607 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:50:09 crc kubenswrapper[4945]: I1206 07:50:09.212145 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"aa5f4b34937f1604b1e9b5f23ae329076fed2880fd3b1678bd6c68cac2d8ae32"} Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.211544 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s42mb"] Dec 06 07:50:53 crc kubenswrapper[4945]: E1206 07:50:53.212356 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20eed229-33e8-4bf8-9620-c797dfab6350" containerName="extract-utilities" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.212373 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="20eed229-33e8-4bf8-9620-c797dfab6350" containerName="extract-utilities" Dec 06 07:50:53 crc kubenswrapper[4945]: E1206 07:50:53.212392 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20eed229-33e8-4bf8-9620-c797dfab6350" containerName="registry-server" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.212399 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="20eed229-33e8-4bf8-9620-c797dfab6350" containerName="registry-server" Dec 06 07:50:53 crc kubenswrapper[4945]: E1206 07:50:53.212415 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20eed229-33e8-4bf8-9620-c797dfab6350" containerName="extract-content" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.212423 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="20eed229-33e8-4bf8-9620-c797dfab6350" containerName="extract-content" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.212603 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="20eed229-33e8-4bf8-9620-c797dfab6350" containerName="registry-server" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.213783 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.222625 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s42mb"] Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.234811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-utilities\") pod \"community-operators-s42mb\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.234961 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-catalog-content\") pod \"community-operators-s42mb\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.235227 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv47g\" (UniqueName: \"kubernetes.io/projected/37113794-2772-4548-8655-a81bca001bbf-kube-api-access-xv47g\") pod \"community-operators-s42mb\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.336497 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv47g\" (UniqueName: \"kubernetes.io/projected/37113794-2772-4548-8655-a81bca001bbf-kube-api-access-xv47g\") pod \"community-operators-s42mb\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.336821 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-utilities\") pod \"community-operators-s42mb\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.336953 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-catalog-content\") pod \"community-operators-s42mb\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.337449 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-catalog-content\") pod \"community-operators-s42mb\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.337467 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-utilities\") pod \"community-operators-s42mb\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.364019 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv47g\" (UniqueName: \"kubernetes.io/projected/37113794-2772-4548-8655-a81bca001bbf-kube-api-access-xv47g\") pod \"community-operators-s42mb\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.412247 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lh46q"] Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.414193 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.438855 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vv2p\" (UniqueName: \"kubernetes.io/projected/96846e1e-2bba-446e-adbd-2862262ad287-kube-api-access-2vv2p\") pod \"redhat-marketplace-lh46q\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.438930 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-catalog-content\") pod \"redhat-marketplace-lh46q\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.438964 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-utilities\") pod \"redhat-marketplace-lh46q\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.460839 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lh46q"] Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.535688 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.540171 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vv2p\" (UniqueName: \"kubernetes.io/projected/96846e1e-2bba-446e-adbd-2862262ad287-kube-api-access-2vv2p\") pod \"redhat-marketplace-lh46q\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.540227 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-catalog-content\") pod \"redhat-marketplace-lh46q\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.540246 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-utilities\") pod \"redhat-marketplace-lh46q\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.540758 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-utilities\") pod \"redhat-marketplace-lh46q\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.540950 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-catalog-content\") pod \"redhat-marketplace-lh46q\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.562713 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vv2p\" (UniqueName: \"kubernetes.io/projected/96846e1e-2bba-446e-adbd-2862262ad287-kube-api-access-2vv2p\") pod \"redhat-marketplace-lh46q\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:53 crc kubenswrapper[4945]: I1206 07:50:53.759624 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:50:54 crc kubenswrapper[4945]: I1206 07:50:54.058774 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s42mb"] Dec 06 07:50:54 crc kubenswrapper[4945]: I1206 07:50:54.204671 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lh46q"] Dec 06 07:50:54 crc kubenswrapper[4945]: W1206 07:50:54.207874 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96846e1e_2bba_446e_adbd_2862262ad287.slice/crio-ee42382cfe66650a8ee2c5cb275fb1e59430c2a26eebaf4c1911d4a7d31dcb6c WatchSource:0}: Error finding container ee42382cfe66650a8ee2c5cb275fb1e59430c2a26eebaf4c1911d4a7d31dcb6c: Status 404 returned error can't find the container with id ee42382cfe66650a8ee2c5cb275fb1e59430c2a26eebaf4c1911d4a7d31dcb6c Dec 06 07:50:54 crc kubenswrapper[4945]: I1206 07:50:54.513968 4945 generic.go:334] "Generic (PLEG): container finished" podID="96846e1e-2bba-446e-adbd-2862262ad287" containerID="83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1" exitCode=0 Dec 06 07:50:54 crc kubenswrapper[4945]: I1206 07:50:54.514047 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lh46q" event={"ID":"96846e1e-2bba-446e-adbd-2862262ad287","Type":"ContainerDied","Data":"83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1"} Dec 06 07:50:54 crc kubenswrapper[4945]: I1206 07:50:54.515085 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lh46q" event={"ID":"96846e1e-2bba-446e-adbd-2862262ad287","Type":"ContainerStarted","Data":"ee42382cfe66650a8ee2c5cb275fb1e59430c2a26eebaf4c1911d4a7d31dcb6c"} Dec 06 07:50:54 crc kubenswrapper[4945]: I1206 07:50:54.517264 4945 generic.go:334] "Generic (PLEG): container finished" podID="37113794-2772-4548-8655-a81bca001bbf" containerID="3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec" exitCode=0 Dec 06 07:50:54 crc kubenswrapper[4945]: I1206 07:50:54.517350 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s42mb" event={"ID":"37113794-2772-4548-8655-a81bca001bbf","Type":"ContainerDied","Data":"3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec"} Dec 06 07:50:54 crc kubenswrapper[4945]: I1206 07:50:54.517383 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s42mb" event={"ID":"37113794-2772-4548-8655-a81bca001bbf","Type":"ContainerStarted","Data":"7c111285bcfca1a714eb79f842b442ed64a314845192d3517d61a0cade75298c"} Dec 06 07:50:58 crc kubenswrapper[4945]: I1206 07:50:58.551434 4945 generic.go:334] "Generic (PLEG): container finished" podID="96846e1e-2bba-446e-adbd-2862262ad287" containerID="56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06" exitCode=0 Dec 06 07:50:58 crc kubenswrapper[4945]: I1206 07:50:58.551488 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lh46q" event={"ID":"96846e1e-2bba-446e-adbd-2862262ad287","Type":"ContainerDied","Data":"56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06"} Dec 06 07:50:58 crc kubenswrapper[4945]: I1206 07:50:58.555726 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s42mb" event={"ID":"37113794-2772-4548-8655-a81bca001bbf","Type":"ContainerStarted","Data":"67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5"} Dec 06 07:50:59 crc kubenswrapper[4945]: I1206 07:50:59.564093 4945 generic.go:334] "Generic (PLEG): container finished" podID="37113794-2772-4548-8655-a81bca001bbf" containerID="67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5" exitCode=0 Dec 06 07:50:59 crc kubenswrapper[4945]: I1206 07:50:59.564155 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s42mb" event={"ID":"37113794-2772-4548-8655-a81bca001bbf","Type":"ContainerDied","Data":"67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5"} Dec 06 07:51:02 crc kubenswrapper[4945]: I1206 07:51:02.590450 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lh46q" event={"ID":"96846e1e-2bba-446e-adbd-2862262ad287","Type":"ContainerStarted","Data":"550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301"} Dec 06 07:51:02 crc kubenswrapper[4945]: I1206 07:51:02.593045 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s42mb" event={"ID":"37113794-2772-4548-8655-a81bca001bbf","Type":"ContainerStarted","Data":"0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d"} Dec 06 07:51:02 crc kubenswrapper[4945]: I1206 07:51:02.634572 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s42mb" podStartSLOduration=2.220569208 podStartE2EDuration="9.634553616s" podCreationTimestamp="2025-12-06 07:50:53 +0000 UTC" firstStartedPulling="2025-12-06 07:50:54.5185992 +0000 UTC m=+3467.973460244" lastFinishedPulling="2025-12-06 07:51:01.932583608 +0000 UTC m=+3475.387444652" observedRunningTime="2025-12-06 07:51:02.631643279 +0000 UTC m=+3476.086504323" watchObservedRunningTime="2025-12-06 07:51:02.634553616 +0000 UTC m=+3476.089414660" Dec 06 07:51:02 crc kubenswrapper[4945]: I1206 07:51:02.635814 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lh46q" podStartSLOduration=2.375324362 podStartE2EDuration="9.63580677s" podCreationTimestamp="2025-12-06 07:50:53 +0000 UTC" firstStartedPulling="2025-12-06 07:50:54.515523658 +0000 UTC m=+3467.970384702" lastFinishedPulling="2025-12-06 07:51:01.776006056 +0000 UTC m=+3475.230867110" observedRunningTime="2025-12-06 07:51:02.61595093 +0000 UTC m=+3476.070811974" watchObservedRunningTime="2025-12-06 07:51:02.63580677 +0000 UTC m=+3476.090667814" Dec 06 07:51:03 crc kubenswrapper[4945]: I1206 07:51:03.536384 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:51:03 crc kubenswrapper[4945]: I1206 07:51:03.536685 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:51:03 crc kubenswrapper[4945]: I1206 07:51:03.759884 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:51:03 crc kubenswrapper[4945]: I1206 07:51:03.759934 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:51:03 crc kubenswrapper[4945]: I1206 07:51:03.801088 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:51:04 crc kubenswrapper[4945]: I1206 07:51:04.580644 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-s42mb" podUID="37113794-2772-4548-8655-a81bca001bbf" containerName="registry-server" probeResult="failure" output=< Dec 06 07:51:04 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 07:51:04 crc kubenswrapper[4945]: > Dec 06 07:51:13 crc kubenswrapper[4945]: I1206 07:51:13.586509 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:51:13 crc kubenswrapper[4945]: I1206 07:51:13.648007 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:51:13 crc kubenswrapper[4945]: I1206 07:51:13.807763 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:51:13 crc kubenswrapper[4945]: I1206 07:51:13.822334 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s42mb"] Dec 06 07:51:14 crc kubenswrapper[4945]: I1206 07:51:14.678663 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s42mb" podUID="37113794-2772-4548-8655-a81bca001bbf" containerName="registry-server" containerID="cri-o://0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d" gracePeriod=2 Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.159767 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.248392 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-utilities\") pod \"37113794-2772-4548-8655-a81bca001bbf\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.248493 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-catalog-content\") pod \"37113794-2772-4548-8655-a81bca001bbf\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.248568 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv47g\" (UniqueName: \"kubernetes.io/projected/37113794-2772-4548-8655-a81bca001bbf-kube-api-access-xv47g\") pod \"37113794-2772-4548-8655-a81bca001bbf\" (UID: \"37113794-2772-4548-8655-a81bca001bbf\") " Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.249538 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-utilities" (OuterVolumeSpecName: "utilities") pod "37113794-2772-4548-8655-a81bca001bbf" (UID: "37113794-2772-4548-8655-a81bca001bbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.253906 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37113794-2772-4548-8655-a81bca001bbf-kube-api-access-xv47g" (OuterVolumeSpecName: "kube-api-access-xv47g") pod "37113794-2772-4548-8655-a81bca001bbf" (UID: "37113794-2772-4548-8655-a81bca001bbf"). InnerVolumeSpecName "kube-api-access-xv47g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.341518 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37113794-2772-4548-8655-a81bca001bbf" (UID: "37113794-2772-4548-8655-a81bca001bbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.350253 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.350302 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37113794-2772-4548-8655-a81bca001bbf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.350316 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xv47g\" (UniqueName: \"kubernetes.io/projected/37113794-2772-4548-8655-a81bca001bbf-kube-api-access-xv47g\") on node \"crc\" DevicePath \"\"" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.688889 4945 generic.go:334] "Generic (PLEG): container finished" podID="37113794-2772-4548-8655-a81bca001bbf" containerID="0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d" exitCode=0 Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.688968 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s42mb" event={"ID":"37113794-2772-4548-8655-a81bca001bbf","Type":"ContainerDied","Data":"0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d"} Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.689017 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s42mb" event={"ID":"37113794-2772-4548-8655-a81bca001bbf","Type":"ContainerDied","Data":"7c111285bcfca1a714eb79f842b442ed64a314845192d3517d61a0cade75298c"} Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.689057 4945 scope.go:117] "RemoveContainer" containerID="0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.689107 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s42mb" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.707947 4945 scope.go:117] "RemoveContainer" containerID="67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.722051 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s42mb"] Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.727529 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s42mb"] Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.738869 4945 scope.go:117] "RemoveContainer" containerID="3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.759574 4945 scope.go:117] "RemoveContainer" containerID="0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d" Dec 06 07:51:15 crc kubenswrapper[4945]: E1206 07:51:15.759893 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d\": container with ID starting with 0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d not found: ID does not exist" containerID="0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.759937 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d"} err="failed to get container status \"0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d\": rpc error: code = NotFound desc = could not find container \"0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d\": container with ID starting with 0c9be2695fa0dca2516ed968ad55b2c9ac5112b368d91cc311d7f8c1ad8a9d0d not found: ID does not exist" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.759963 4945 scope.go:117] "RemoveContainer" containerID="67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5" Dec 06 07:51:15 crc kubenswrapper[4945]: E1206 07:51:15.760238 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5\": container with ID starting with 67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5 not found: ID does not exist" containerID="67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.760264 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5"} err="failed to get container status \"67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5\": rpc error: code = NotFound desc = could not find container \"67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5\": container with ID starting with 67df62993d9cbee18e925d6fe2479bae03b0a4c079a89ea07ff275a6036980b5 not found: ID does not exist" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.760301 4945 scope.go:117] "RemoveContainer" containerID="3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec" Dec 06 07:51:15 crc kubenswrapper[4945]: E1206 07:51:15.760499 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec\": container with ID starting with 3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec not found: ID does not exist" containerID="3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec" Dec 06 07:51:15 crc kubenswrapper[4945]: I1206 07:51:15.760517 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec"} err="failed to get container status \"3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec\": rpc error: code = NotFound desc = could not find container \"3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec\": container with ID starting with 3d4e4b51e4bff5d5b15c9679b6fb5938940c83b08220ae916e0cc76a1483b6ec not found: ID does not exist" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.221553 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lh46q"] Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.221823 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lh46q" podUID="96846e1e-2bba-446e-adbd-2862262ad287" containerName="registry-server" containerID="cri-o://550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301" gracePeriod=2 Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.624465 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.668478 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-utilities\") pod \"96846e1e-2bba-446e-adbd-2862262ad287\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.668538 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-catalog-content\") pod \"96846e1e-2bba-446e-adbd-2862262ad287\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.668640 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vv2p\" (UniqueName: \"kubernetes.io/projected/96846e1e-2bba-446e-adbd-2862262ad287-kube-api-access-2vv2p\") pod \"96846e1e-2bba-446e-adbd-2862262ad287\" (UID: \"96846e1e-2bba-446e-adbd-2862262ad287\") " Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.670038 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-utilities" (OuterVolumeSpecName: "utilities") pod "96846e1e-2bba-446e-adbd-2862262ad287" (UID: "96846e1e-2bba-446e-adbd-2862262ad287"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.677422 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96846e1e-2bba-446e-adbd-2862262ad287-kube-api-access-2vv2p" (OuterVolumeSpecName: "kube-api-access-2vv2p") pod "96846e1e-2bba-446e-adbd-2862262ad287" (UID: "96846e1e-2bba-446e-adbd-2862262ad287"). InnerVolumeSpecName "kube-api-access-2vv2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.694372 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96846e1e-2bba-446e-adbd-2862262ad287" (UID: "96846e1e-2bba-446e-adbd-2862262ad287"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.701487 4945 generic.go:334] "Generic (PLEG): container finished" podID="96846e1e-2bba-446e-adbd-2862262ad287" containerID="550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301" exitCode=0 Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.701621 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lh46q" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.701544 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lh46q" event={"ID":"96846e1e-2bba-446e-adbd-2862262ad287","Type":"ContainerDied","Data":"550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301"} Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.701821 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lh46q" event={"ID":"96846e1e-2bba-446e-adbd-2862262ad287","Type":"ContainerDied","Data":"ee42382cfe66650a8ee2c5cb275fb1e59430c2a26eebaf4c1911d4a7d31dcb6c"} Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.701846 4945 scope.go:117] "RemoveContainer" containerID="550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.738694 4945 scope.go:117] "RemoveContainer" containerID="56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.743539 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lh46q"] Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.751219 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lh46q"] Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.768995 4945 scope.go:117] "RemoveContainer" containerID="83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.769835 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vv2p\" (UniqueName: \"kubernetes.io/projected/96846e1e-2bba-446e-adbd-2862262ad287-kube-api-access-2vv2p\") on node \"crc\" DevicePath \"\"" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.769867 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.769880 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96846e1e-2bba-446e-adbd-2862262ad287-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.786238 4945 scope.go:117] "RemoveContainer" containerID="550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301" Dec 06 07:51:16 crc kubenswrapper[4945]: E1206 07:51:16.786787 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301\": container with ID starting with 550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301 not found: ID does not exist" containerID="550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.786834 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301"} err="failed to get container status \"550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301\": rpc error: code = NotFound desc = could not find container \"550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301\": container with ID starting with 550a82a05879cbf4a9182914a54f984c002d22dab7211729df1be37e1ec41301 not found: ID does not exist" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.786868 4945 scope.go:117] "RemoveContainer" containerID="56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06" Dec 06 07:51:16 crc kubenswrapper[4945]: E1206 07:51:16.787248 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06\": container with ID starting with 56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06 not found: ID does not exist" containerID="56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.787326 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06"} err="failed to get container status \"56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06\": rpc error: code = NotFound desc = could not find container \"56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06\": container with ID starting with 56e29f39b5964f77dd7b5646d5db87b5ec00c63b45937c7f59d0318d1533bb06 not found: ID does not exist" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.787348 4945 scope.go:117] "RemoveContainer" containerID="83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1" Dec 06 07:51:16 crc kubenswrapper[4945]: E1206 07:51:16.787622 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1\": container with ID starting with 83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1 not found: ID does not exist" containerID="83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.787650 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1"} err="failed to get container status \"83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1\": rpc error: code = NotFound desc = could not find container \"83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1\": container with ID starting with 83991ad01194fe735e79e4e64c41a59a146af66f68f5a6070f24dea7e5d520f1 not found: ID does not exist" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.966153 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37113794-2772-4548-8655-a81bca001bbf" path="/var/lib/kubelet/pods/37113794-2772-4548-8655-a81bca001bbf/volumes" Dec 06 07:51:16 crc kubenswrapper[4945]: I1206 07:51:16.967051 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96846e1e-2bba-446e-adbd-2862262ad287" path="/var/lib/kubelet/pods/96846e1e-2bba-446e-adbd-2862262ad287/volumes" Dec 06 07:52:38 crc kubenswrapper[4945]: I1206 07:52:38.795432 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:52:38 crc kubenswrapper[4945]: I1206 07:52:38.796312 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:53:08 crc kubenswrapper[4945]: I1206 07:53:08.795880 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:53:08 crc kubenswrapper[4945]: I1206 07:53:08.796513 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.428618 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x56gn"] Dec 06 07:53:33 crc kubenswrapper[4945]: E1206 07:53:33.429492 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37113794-2772-4548-8655-a81bca001bbf" containerName="extract-content" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.429509 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="37113794-2772-4548-8655-a81bca001bbf" containerName="extract-content" Dec 06 07:53:33 crc kubenswrapper[4945]: E1206 07:53:33.429526 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37113794-2772-4548-8655-a81bca001bbf" containerName="registry-server" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.429532 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="37113794-2772-4548-8655-a81bca001bbf" containerName="registry-server" Dec 06 07:53:33 crc kubenswrapper[4945]: E1206 07:53:33.429548 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37113794-2772-4548-8655-a81bca001bbf" containerName="extract-utilities" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.429554 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="37113794-2772-4548-8655-a81bca001bbf" containerName="extract-utilities" Dec 06 07:53:33 crc kubenswrapper[4945]: E1206 07:53:33.429585 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96846e1e-2bba-446e-adbd-2862262ad287" containerName="registry-server" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.429598 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="96846e1e-2bba-446e-adbd-2862262ad287" containerName="registry-server" Dec 06 07:53:33 crc kubenswrapper[4945]: E1206 07:53:33.429616 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96846e1e-2bba-446e-adbd-2862262ad287" containerName="extract-content" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.429624 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="96846e1e-2bba-446e-adbd-2862262ad287" containerName="extract-content" Dec 06 07:53:33 crc kubenswrapper[4945]: E1206 07:53:33.429633 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96846e1e-2bba-446e-adbd-2862262ad287" containerName="extract-utilities" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.429639 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="96846e1e-2bba-446e-adbd-2862262ad287" containerName="extract-utilities" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.429783 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="96846e1e-2bba-446e-adbd-2862262ad287" containerName="registry-server" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.429821 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="37113794-2772-4548-8655-a81bca001bbf" containerName="registry-server" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.430946 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.448168 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x56gn"] Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.576501 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct8c5\" (UniqueName: \"kubernetes.io/projected/b516e13c-1bf9-4590-93bc-b6c063d0d940-kube-api-access-ct8c5\") pod \"redhat-operators-x56gn\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.576967 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-catalog-content\") pod \"redhat-operators-x56gn\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.577223 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-utilities\") pod \"redhat-operators-x56gn\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.679746 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct8c5\" (UniqueName: \"kubernetes.io/projected/b516e13c-1bf9-4590-93bc-b6c063d0d940-kube-api-access-ct8c5\") pod \"redhat-operators-x56gn\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.679869 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-catalog-content\") pod \"redhat-operators-x56gn\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.679966 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-utilities\") pod \"redhat-operators-x56gn\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.680560 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-catalog-content\") pod \"redhat-operators-x56gn\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.680585 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-utilities\") pod \"redhat-operators-x56gn\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.709304 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct8c5\" (UniqueName: \"kubernetes.io/projected/b516e13c-1bf9-4590-93bc-b6c063d0d940-kube-api-access-ct8c5\") pod \"redhat-operators-x56gn\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:33 crc kubenswrapper[4945]: I1206 07:53:33.753077 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:34 crc kubenswrapper[4945]: I1206 07:53:34.126322 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x56gn"] Dec 06 07:53:34 crc kubenswrapper[4945]: W1206 07:53:34.134304 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb516e13c_1bf9_4590_93bc_b6c063d0d940.slice/crio-cd1ca364743ccee3b7f8cb28ccf9b966de92267b78ccbf3ed7ea3fbd5d2fb9e8 WatchSource:0}: Error finding container cd1ca364743ccee3b7f8cb28ccf9b966de92267b78ccbf3ed7ea3fbd5d2fb9e8: Status 404 returned error can't find the container with id cd1ca364743ccee3b7f8cb28ccf9b966de92267b78ccbf3ed7ea3fbd5d2fb9e8 Dec 06 07:53:34 crc kubenswrapper[4945]: I1206 07:53:34.775692 4945 generic.go:334] "Generic (PLEG): container finished" podID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerID="012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45" exitCode=0 Dec 06 07:53:34 crc kubenswrapper[4945]: I1206 07:53:34.775763 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56gn" event={"ID":"b516e13c-1bf9-4590-93bc-b6c063d0d940","Type":"ContainerDied","Data":"012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45"} Dec 06 07:53:34 crc kubenswrapper[4945]: I1206 07:53:34.775859 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56gn" event={"ID":"b516e13c-1bf9-4590-93bc-b6c063d0d940","Type":"ContainerStarted","Data":"cd1ca364743ccee3b7f8cb28ccf9b966de92267b78ccbf3ed7ea3fbd5d2fb9e8"} Dec 06 07:53:34 crc kubenswrapper[4945]: I1206 07:53:34.778809 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:53:35 crc kubenswrapper[4945]: I1206 07:53:35.783892 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56gn" event={"ID":"b516e13c-1bf9-4590-93bc-b6c063d0d940","Type":"ContainerStarted","Data":"072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c"} Dec 06 07:53:36 crc kubenswrapper[4945]: I1206 07:53:36.793068 4945 generic.go:334] "Generic (PLEG): container finished" podID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerID="072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c" exitCode=0 Dec 06 07:53:36 crc kubenswrapper[4945]: I1206 07:53:36.793164 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56gn" event={"ID":"b516e13c-1bf9-4590-93bc-b6c063d0d940","Type":"ContainerDied","Data":"072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c"} Dec 06 07:53:38 crc kubenswrapper[4945]: I1206 07:53:38.796035 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:53:38 crc kubenswrapper[4945]: I1206 07:53:38.796615 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:53:38 crc kubenswrapper[4945]: I1206 07:53:38.796675 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:53:38 crc kubenswrapper[4945]: I1206 07:53:38.797486 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa5f4b34937f1604b1e9b5f23ae329076fed2880fd3b1678bd6c68cac2d8ae32"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:53:38 crc kubenswrapper[4945]: I1206 07:53:38.797564 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://aa5f4b34937f1604b1e9b5f23ae329076fed2880fd3b1678bd6c68cac2d8ae32" gracePeriod=600 Dec 06 07:53:38 crc kubenswrapper[4945]: I1206 07:53:38.812643 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56gn" event={"ID":"b516e13c-1bf9-4590-93bc-b6c063d0d940","Type":"ContainerStarted","Data":"4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768"} Dec 06 07:53:38 crc kubenswrapper[4945]: I1206 07:53:38.834685 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x56gn" podStartSLOduration=2.333875392 podStartE2EDuration="5.834668244s" podCreationTimestamp="2025-12-06 07:53:33 +0000 UTC" firstStartedPulling="2025-12-06 07:53:34.778518249 +0000 UTC m=+3628.233379293" lastFinishedPulling="2025-12-06 07:53:38.279311081 +0000 UTC m=+3631.734172145" observedRunningTime="2025-12-06 07:53:38.830749409 +0000 UTC m=+3632.285610483" watchObservedRunningTime="2025-12-06 07:53:38.834668244 +0000 UTC m=+3632.289529278" Dec 06 07:53:39 crc kubenswrapper[4945]: I1206 07:53:39.823954 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="aa5f4b34937f1604b1e9b5f23ae329076fed2880fd3b1678bd6c68cac2d8ae32" exitCode=0 Dec 06 07:53:39 crc kubenswrapper[4945]: I1206 07:53:39.824049 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"aa5f4b34937f1604b1e9b5f23ae329076fed2880fd3b1678bd6c68cac2d8ae32"} Dec 06 07:53:39 crc kubenswrapper[4945]: I1206 07:53:39.824546 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8"} Dec 06 07:53:39 crc kubenswrapper[4945]: I1206 07:53:39.824574 4945 scope.go:117] "RemoveContainer" containerID="e3273ba3b25bef704fa8c162172ba0d90d9be61622c2760f4defc13ba06cd80e" Dec 06 07:53:43 crc kubenswrapper[4945]: I1206 07:53:43.754886 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:43 crc kubenswrapper[4945]: I1206 07:53:43.757656 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:43 crc kubenswrapper[4945]: I1206 07:53:43.816601 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:43 crc kubenswrapper[4945]: I1206 07:53:43.899001 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:44 crc kubenswrapper[4945]: I1206 07:53:44.053947 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x56gn"] Dec 06 07:53:45 crc kubenswrapper[4945]: I1206 07:53:45.871881 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x56gn" podUID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerName="registry-server" containerID="cri-o://4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768" gracePeriod=2 Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.243905 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.375702 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct8c5\" (UniqueName: \"kubernetes.io/projected/b516e13c-1bf9-4590-93bc-b6c063d0d940-kube-api-access-ct8c5\") pod \"b516e13c-1bf9-4590-93bc-b6c063d0d940\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.375772 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-catalog-content\") pod \"b516e13c-1bf9-4590-93bc-b6c063d0d940\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.376055 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-utilities\") pod \"b516e13c-1bf9-4590-93bc-b6c063d0d940\" (UID: \"b516e13c-1bf9-4590-93bc-b6c063d0d940\") " Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.377536 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-utilities" (OuterVolumeSpecName: "utilities") pod "b516e13c-1bf9-4590-93bc-b6c063d0d940" (UID: "b516e13c-1bf9-4590-93bc-b6c063d0d940"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.382852 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b516e13c-1bf9-4590-93bc-b6c063d0d940-kube-api-access-ct8c5" (OuterVolumeSpecName: "kube-api-access-ct8c5") pod "b516e13c-1bf9-4590-93bc-b6c063d0d940" (UID: "b516e13c-1bf9-4590-93bc-b6c063d0d940"). InnerVolumeSpecName "kube-api-access-ct8c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.479188 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.479235 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct8c5\" (UniqueName: \"kubernetes.io/projected/b516e13c-1bf9-4590-93bc-b6c063d0d940-kube-api-access-ct8c5\") on node \"crc\" DevicePath \"\"" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.880023 4945 generic.go:334] "Generic (PLEG): container finished" podID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerID="4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768" exitCode=0 Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.880064 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56gn" event={"ID":"b516e13c-1bf9-4590-93bc-b6c063d0d940","Type":"ContainerDied","Data":"4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768"} Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.880099 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x56gn" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.880361 4945 scope.go:117] "RemoveContainer" containerID="4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.880348 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56gn" event={"ID":"b516e13c-1bf9-4590-93bc-b6c063d0d940","Type":"ContainerDied","Data":"cd1ca364743ccee3b7f8cb28ccf9b966de92267b78ccbf3ed7ea3fbd5d2fb9e8"} Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.897842 4945 scope.go:117] "RemoveContainer" containerID="072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.921503 4945 scope.go:117] "RemoveContainer" containerID="012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.942415 4945 scope.go:117] "RemoveContainer" containerID="4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768" Dec 06 07:53:46 crc kubenswrapper[4945]: E1206 07:53:46.943263 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768\": container with ID starting with 4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768 not found: ID does not exist" containerID="4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.943342 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768"} err="failed to get container status \"4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768\": rpc error: code = NotFound desc = could not find container \"4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768\": container with ID starting with 4cc874112ac490cf286690cb1ed5941b902d4ba71764c62aceddbc524fee6768 not found: ID does not exist" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.943376 4945 scope.go:117] "RemoveContainer" containerID="072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c" Dec 06 07:53:46 crc kubenswrapper[4945]: E1206 07:53:46.943838 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c\": container with ID starting with 072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c not found: ID does not exist" containerID="072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.943886 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c"} err="failed to get container status \"072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c\": rpc error: code = NotFound desc = could not find container \"072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c\": container with ID starting with 072d82658b94b5149210f886caf2f7bb1104299e2bd9e96523d8c8eec56a478c not found: ID does not exist" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.943918 4945 scope.go:117] "RemoveContainer" containerID="012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45" Dec 06 07:53:46 crc kubenswrapper[4945]: E1206 07:53:46.944166 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45\": container with ID starting with 012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45 not found: ID does not exist" containerID="012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45" Dec 06 07:53:46 crc kubenswrapper[4945]: I1206 07:53:46.944202 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45"} err="failed to get container status \"012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45\": rpc error: code = NotFound desc = could not find container \"012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45\": container with ID starting with 012f637d6a054dddda586fedfc6d092fc09ad612a9d83b971cd8467db8c7ac45 not found: ID does not exist" Dec 06 07:53:47 crc kubenswrapper[4945]: I1206 07:53:47.365878 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b516e13c-1bf9-4590-93bc-b6c063d0d940" (UID: "b516e13c-1bf9-4590-93bc-b6c063d0d940"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:53:47 crc kubenswrapper[4945]: I1206 07:53:47.393534 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b516e13c-1bf9-4590-93bc-b6c063d0d940-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:53:47 crc kubenswrapper[4945]: I1206 07:53:47.518160 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x56gn"] Dec 06 07:53:47 crc kubenswrapper[4945]: I1206 07:53:47.525447 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x56gn"] Dec 06 07:53:48 crc kubenswrapper[4945]: I1206 07:53:48.965972 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b516e13c-1bf9-4590-93bc-b6c063d0d940" path="/var/lib/kubelet/pods/b516e13c-1bf9-4590-93bc-b6c063d0d940/volumes" Dec 06 07:56:08 crc kubenswrapper[4945]: I1206 07:56:08.795637 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:56:08 crc kubenswrapper[4945]: I1206 07:56:08.796239 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:56:38 crc kubenswrapper[4945]: I1206 07:56:38.796324 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:56:38 crc kubenswrapper[4945]: I1206 07:56:38.797062 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:57:08 crc kubenswrapper[4945]: I1206 07:57:08.795345 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:57:08 crc kubenswrapper[4945]: I1206 07:57:08.795993 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:57:08 crc kubenswrapper[4945]: I1206 07:57:08.796043 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 07:57:08 crc kubenswrapper[4945]: I1206 07:57:08.796540 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:57:08 crc kubenswrapper[4945]: I1206 07:57:08.796593 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" gracePeriod=600 Dec 06 07:57:08 crc kubenswrapper[4945]: E1206 07:57:08.922840 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:57:09 crc kubenswrapper[4945]: I1206 07:57:09.531838 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" exitCode=0 Dec 06 07:57:09 crc kubenswrapper[4945]: I1206 07:57:09.531917 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8"} Dec 06 07:57:09 crc kubenswrapper[4945]: I1206 07:57:09.532153 4945 scope.go:117] "RemoveContainer" containerID="aa5f4b34937f1604b1e9b5f23ae329076fed2880fd3b1678bd6c68cac2d8ae32" Dec 06 07:57:09 crc kubenswrapper[4945]: I1206 07:57:09.532643 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:57:09 crc kubenswrapper[4945]: E1206 07:57:09.532878 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:57:23 crc kubenswrapper[4945]: I1206 07:57:23.952935 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:57:23 crc kubenswrapper[4945]: E1206 07:57:23.953696 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:57:37 crc kubenswrapper[4945]: I1206 07:57:37.953928 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:57:37 crc kubenswrapper[4945]: E1206 07:57:37.954759 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:57:49 crc kubenswrapper[4945]: I1206 07:57:49.953306 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:57:49 crc kubenswrapper[4945]: E1206 07:57:49.954193 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:58:02 crc kubenswrapper[4945]: I1206 07:58:02.953189 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:58:02 crc kubenswrapper[4945]: E1206 07:58:02.953940 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:58:13 crc kubenswrapper[4945]: I1206 07:58:13.953766 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:58:13 crc kubenswrapper[4945]: E1206 07:58:13.954704 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:58:24 crc kubenswrapper[4945]: I1206 07:58:24.954042 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:58:24 crc kubenswrapper[4945]: E1206 07:58:24.954862 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.476484 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f25tn"] Dec 06 07:58:36 crc kubenswrapper[4945]: E1206 07:58:36.477479 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerName="registry-server" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.477495 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerName="registry-server" Dec 06 07:58:36 crc kubenswrapper[4945]: E1206 07:58:36.477513 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerName="extract-content" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.477520 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerName="extract-content" Dec 06 07:58:36 crc kubenswrapper[4945]: E1206 07:58:36.477539 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerName="extract-utilities" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.477547 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerName="extract-utilities" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.477733 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b516e13c-1bf9-4590-93bc-b6c063d0d940" containerName="registry-server" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.479040 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.502953 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f25tn"] Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.590643 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7xs7\" (UniqueName: \"kubernetes.io/projected/7311010d-ec15-49a0-9f4e-4af13b79e7e2-kube-api-access-b7xs7\") pod \"certified-operators-f25tn\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.590696 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-utilities\") pod \"certified-operators-f25tn\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.590733 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-catalog-content\") pod \"certified-operators-f25tn\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.692579 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7xs7\" (UniqueName: \"kubernetes.io/projected/7311010d-ec15-49a0-9f4e-4af13b79e7e2-kube-api-access-b7xs7\") pod \"certified-operators-f25tn\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.692633 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-utilities\") pod \"certified-operators-f25tn\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.692670 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-catalog-content\") pod \"certified-operators-f25tn\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.693219 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-catalog-content\") pod \"certified-operators-f25tn\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.693260 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-utilities\") pod \"certified-operators-f25tn\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.713479 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7xs7\" (UniqueName: \"kubernetes.io/projected/7311010d-ec15-49a0-9f4e-4af13b79e7e2-kube-api-access-b7xs7\") pod \"certified-operators-f25tn\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:36 crc kubenswrapper[4945]: I1206 07:58:36.804463 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:37 crc kubenswrapper[4945]: I1206 07:58:37.340054 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f25tn"] Dec 06 07:58:37 crc kubenswrapper[4945]: I1206 07:58:37.963571 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:58:37 crc kubenswrapper[4945]: E1206 07:58:37.964892 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:58:38 crc kubenswrapper[4945]: I1206 07:58:38.167330 4945 generic.go:334] "Generic (PLEG): container finished" podID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerID="dd9804a5af3d7b40ee696903f8961953ccab86b1dcbdd502e812ac1fcab809e6" exitCode=0 Dec 06 07:58:38 crc kubenswrapper[4945]: I1206 07:58:38.167399 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f25tn" event={"ID":"7311010d-ec15-49a0-9f4e-4af13b79e7e2","Type":"ContainerDied","Data":"dd9804a5af3d7b40ee696903f8961953ccab86b1dcbdd502e812ac1fcab809e6"} Dec 06 07:58:38 crc kubenswrapper[4945]: I1206 07:58:38.167482 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f25tn" event={"ID":"7311010d-ec15-49a0-9f4e-4af13b79e7e2","Type":"ContainerStarted","Data":"c67002d8c892378453eecbccafb31dee46b06e2ed68a8f843397c0689b189bf5"} Dec 06 07:58:38 crc kubenswrapper[4945]: I1206 07:58:38.169877 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:58:39 crc kubenswrapper[4945]: I1206 07:58:39.176067 4945 generic.go:334] "Generic (PLEG): container finished" podID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerID="aded43274752ed88e3f0c4428e16abb11077aa30675e88fa8375cbad18cb1f62" exitCode=0 Dec 06 07:58:39 crc kubenswrapper[4945]: I1206 07:58:39.176485 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f25tn" event={"ID":"7311010d-ec15-49a0-9f4e-4af13b79e7e2","Type":"ContainerDied","Data":"aded43274752ed88e3f0c4428e16abb11077aa30675e88fa8375cbad18cb1f62"} Dec 06 07:58:40 crc kubenswrapper[4945]: I1206 07:58:40.187352 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f25tn" event={"ID":"7311010d-ec15-49a0-9f4e-4af13b79e7e2","Type":"ContainerStarted","Data":"0e5b414691fa4ac211ff9a29babaa0649442449f4de07d3756547b70f0c7221f"} Dec 06 07:58:46 crc kubenswrapper[4945]: I1206 07:58:46.805229 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:46 crc kubenswrapper[4945]: I1206 07:58:46.805517 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:46 crc kubenswrapper[4945]: I1206 07:58:46.841600 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:46 crc kubenswrapper[4945]: I1206 07:58:46.861806 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f25tn" podStartSLOduration=9.465121706 podStartE2EDuration="10.861788022s" podCreationTimestamp="2025-12-06 07:58:36 +0000 UTC" firstStartedPulling="2025-12-06 07:58:38.169565811 +0000 UTC m=+3931.624426855" lastFinishedPulling="2025-12-06 07:58:39.566232127 +0000 UTC m=+3933.021093171" observedRunningTime="2025-12-06 07:58:40.212155092 +0000 UTC m=+3933.667016136" watchObservedRunningTime="2025-12-06 07:58:46.861788022 +0000 UTC m=+3940.316649066" Dec 06 07:58:47 crc kubenswrapper[4945]: I1206 07:58:47.307716 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:47 crc kubenswrapper[4945]: I1206 07:58:47.349458 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f25tn"] Dec 06 07:58:49 crc kubenswrapper[4945]: I1206 07:58:49.270107 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f25tn" podUID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerName="registry-server" containerID="cri-o://0e5b414691fa4ac211ff9a29babaa0649442449f4de07d3756547b70f0c7221f" gracePeriod=2 Dec 06 07:58:50 crc kubenswrapper[4945]: I1206 07:58:50.277395 4945 generic.go:334] "Generic (PLEG): container finished" podID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerID="0e5b414691fa4ac211ff9a29babaa0649442449f4de07d3756547b70f0c7221f" exitCode=0 Dec 06 07:58:50 crc kubenswrapper[4945]: I1206 07:58:50.277577 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f25tn" event={"ID":"7311010d-ec15-49a0-9f4e-4af13b79e7e2","Type":"ContainerDied","Data":"0e5b414691fa4ac211ff9a29babaa0649442449f4de07d3756547b70f0c7221f"} Dec 06 07:58:50 crc kubenswrapper[4945]: I1206 07:58:50.782222 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:50 crc kubenswrapper[4945]: I1206 07:58:50.905914 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-utilities\") pod \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " Dec 06 07:58:50 crc kubenswrapper[4945]: I1206 07:58:50.906027 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7xs7\" (UniqueName: \"kubernetes.io/projected/7311010d-ec15-49a0-9f4e-4af13b79e7e2-kube-api-access-b7xs7\") pod \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " Dec 06 07:58:50 crc kubenswrapper[4945]: I1206 07:58:50.906056 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-catalog-content\") pod \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\" (UID: \"7311010d-ec15-49a0-9f4e-4af13b79e7e2\") " Dec 06 07:58:50 crc kubenswrapper[4945]: I1206 07:58:50.907190 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-utilities" (OuterVolumeSpecName: "utilities") pod "7311010d-ec15-49a0-9f4e-4af13b79e7e2" (UID: "7311010d-ec15-49a0-9f4e-4af13b79e7e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:58:50 crc kubenswrapper[4945]: I1206 07:58:50.912463 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7311010d-ec15-49a0-9f4e-4af13b79e7e2-kube-api-access-b7xs7" (OuterVolumeSpecName: "kube-api-access-b7xs7") pod "7311010d-ec15-49a0-9f4e-4af13b79e7e2" (UID: "7311010d-ec15-49a0-9f4e-4af13b79e7e2"). InnerVolumeSpecName "kube-api-access-b7xs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:58:50 crc kubenswrapper[4945]: I1206 07:58:50.957205 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7311010d-ec15-49a0-9f4e-4af13b79e7e2" (UID: "7311010d-ec15-49a0-9f4e-4af13b79e7e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:58:51 crc kubenswrapper[4945]: I1206 07:58:51.007366 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:58:51 crc kubenswrapper[4945]: I1206 07:58:51.007421 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7xs7\" (UniqueName: \"kubernetes.io/projected/7311010d-ec15-49a0-9f4e-4af13b79e7e2-kube-api-access-b7xs7\") on node \"crc\" DevicePath \"\"" Dec 06 07:58:51 crc kubenswrapper[4945]: I1206 07:58:51.007440 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7311010d-ec15-49a0-9f4e-4af13b79e7e2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:58:51 crc kubenswrapper[4945]: I1206 07:58:51.286744 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f25tn" event={"ID":"7311010d-ec15-49a0-9f4e-4af13b79e7e2","Type":"ContainerDied","Data":"c67002d8c892378453eecbccafb31dee46b06e2ed68a8f843397c0689b189bf5"} Dec 06 07:58:51 crc kubenswrapper[4945]: I1206 07:58:51.286797 4945 scope.go:117] "RemoveContainer" containerID="0e5b414691fa4ac211ff9a29babaa0649442449f4de07d3756547b70f0c7221f" Dec 06 07:58:51 crc kubenswrapper[4945]: I1206 07:58:51.286933 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f25tn" Dec 06 07:58:51 crc kubenswrapper[4945]: I1206 07:58:51.314768 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f25tn"] Dec 06 07:58:51 crc kubenswrapper[4945]: I1206 07:58:51.318219 4945 scope.go:117] "RemoveContainer" containerID="aded43274752ed88e3f0c4428e16abb11077aa30675e88fa8375cbad18cb1f62" Dec 06 07:58:51 crc kubenswrapper[4945]: I1206 07:58:51.321005 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f25tn"] Dec 06 07:58:51 crc kubenswrapper[4945]: I1206 07:58:51.339413 4945 scope.go:117] "RemoveContainer" containerID="dd9804a5af3d7b40ee696903f8961953ccab86b1dcbdd502e812ac1fcab809e6" Dec 06 07:58:52 crc kubenswrapper[4945]: I1206 07:58:52.953898 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:58:52 crc kubenswrapper[4945]: E1206 07:58:52.954386 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:58:52 crc kubenswrapper[4945]: I1206 07:58:52.964184 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" path="/var/lib/kubelet/pods/7311010d-ec15-49a0-9f4e-4af13b79e7e2/volumes" Dec 06 07:59:07 crc kubenswrapper[4945]: I1206 07:59:07.953434 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:59:07 crc kubenswrapper[4945]: E1206 07:59:07.954152 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:59:19 crc kubenswrapper[4945]: I1206 07:59:19.953327 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:59:19 crc kubenswrapper[4945]: E1206 07:59:19.953949 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:59:31 crc kubenswrapper[4945]: I1206 07:59:31.953661 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:59:31 crc kubenswrapper[4945]: E1206 07:59:31.954699 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:59:44 crc kubenswrapper[4945]: I1206 07:59:44.953413 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:59:44 crc kubenswrapper[4945]: E1206 07:59:44.954139 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 07:59:59 crc kubenswrapper[4945]: I1206 07:59:59.953921 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 07:59:59 crc kubenswrapper[4945]: E1206 07:59:59.955847 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.182699 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng"] Dec 06 08:00:00 crc kubenswrapper[4945]: E1206 08:00:00.183057 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerName="registry-server" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.183078 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerName="registry-server" Dec 06 08:00:00 crc kubenswrapper[4945]: E1206 08:00:00.183109 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerName="extract-utilities" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.183116 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerName="extract-utilities" Dec 06 08:00:00 crc kubenswrapper[4945]: E1206 08:00:00.183126 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerName="extract-content" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.183131 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerName="extract-content" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.183330 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7311010d-ec15-49a0-9f4e-4af13b79e7e2" containerName="registry-server" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.183935 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.186601 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.186889 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.194053 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng"] Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.263952 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/78f0df6a-316c-4970-b612-3adb2e6cfb45-secret-volume\") pod \"collect-profiles-29416800-d9jng\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.264666 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tptnt\" (UniqueName: \"kubernetes.io/projected/78f0df6a-316c-4970-b612-3adb2e6cfb45-kube-api-access-tptnt\") pod \"collect-profiles-29416800-d9jng\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.264791 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/78f0df6a-316c-4970-b612-3adb2e6cfb45-config-volume\") pod \"collect-profiles-29416800-d9jng\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.366589 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/78f0df6a-316c-4970-b612-3adb2e6cfb45-config-volume\") pod \"collect-profiles-29416800-d9jng\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.366651 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/78f0df6a-316c-4970-b612-3adb2e6cfb45-secret-volume\") pod \"collect-profiles-29416800-d9jng\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.366707 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tptnt\" (UniqueName: \"kubernetes.io/projected/78f0df6a-316c-4970-b612-3adb2e6cfb45-kube-api-access-tptnt\") pod \"collect-profiles-29416800-d9jng\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.368442 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/78f0df6a-316c-4970-b612-3adb2e6cfb45-config-volume\") pod \"collect-profiles-29416800-d9jng\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.375199 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/78f0df6a-316c-4970-b612-3adb2e6cfb45-secret-volume\") pod \"collect-profiles-29416800-d9jng\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.381696 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tptnt\" (UniqueName: \"kubernetes.io/projected/78f0df6a-316c-4970-b612-3adb2e6cfb45-kube-api-access-tptnt\") pod \"collect-profiles-29416800-d9jng\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.509384 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:00 crc kubenswrapper[4945]: I1206 08:00:00.920587 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng"] Dec 06 08:00:01 crc kubenswrapper[4945]: I1206 08:00:01.786462 4945 generic.go:334] "Generic (PLEG): container finished" podID="78f0df6a-316c-4970-b612-3adb2e6cfb45" containerID="0f4d8a56020be620bcf286ab050f74813cdeea9e235df4abb0fcb795d5debe8c" exitCode=0 Dec 06 08:00:01 crc kubenswrapper[4945]: I1206 08:00:01.786509 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" event={"ID":"78f0df6a-316c-4970-b612-3adb2e6cfb45","Type":"ContainerDied","Data":"0f4d8a56020be620bcf286ab050f74813cdeea9e235df4abb0fcb795d5debe8c"} Dec 06 08:00:01 crc kubenswrapper[4945]: I1206 08:00:01.786538 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" event={"ID":"78f0df6a-316c-4970-b612-3adb2e6cfb45","Type":"ContainerStarted","Data":"854d47d4492f0360352fbfff121d08b2ca44a0f93f56cac3618921646816c359"} Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.134610 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.206811 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/78f0df6a-316c-4970-b612-3adb2e6cfb45-config-volume\") pod \"78f0df6a-316c-4970-b612-3adb2e6cfb45\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.206924 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tptnt\" (UniqueName: \"kubernetes.io/projected/78f0df6a-316c-4970-b612-3adb2e6cfb45-kube-api-access-tptnt\") pod \"78f0df6a-316c-4970-b612-3adb2e6cfb45\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.206945 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/78f0df6a-316c-4970-b612-3adb2e6cfb45-secret-volume\") pod \"78f0df6a-316c-4970-b612-3adb2e6cfb45\" (UID: \"78f0df6a-316c-4970-b612-3adb2e6cfb45\") " Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.207686 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78f0df6a-316c-4970-b612-3adb2e6cfb45-config-volume" (OuterVolumeSpecName: "config-volume") pod "78f0df6a-316c-4970-b612-3adb2e6cfb45" (UID: "78f0df6a-316c-4970-b612-3adb2e6cfb45"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.212540 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78f0df6a-316c-4970-b612-3adb2e6cfb45-kube-api-access-tptnt" (OuterVolumeSpecName: "kube-api-access-tptnt") pod "78f0df6a-316c-4970-b612-3adb2e6cfb45" (UID: "78f0df6a-316c-4970-b612-3adb2e6cfb45"). InnerVolumeSpecName "kube-api-access-tptnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.212646 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78f0df6a-316c-4970-b612-3adb2e6cfb45-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "78f0df6a-316c-4970-b612-3adb2e6cfb45" (UID: "78f0df6a-316c-4970-b612-3adb2e6cfb45"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.309395 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/78f0df6a-316c-4970-b612-3adb2e6cfb45-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.309455 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tptnt\" (UniqueName: \"kubernetes.io/projected/78f0df6a-316c-4970-b612-3adb2e6cfb45-kube-api-access-tptnt\") on node \"crc\" DevicePath \"\"" Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.309475 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/78f0df6a-316c-4970-b612-3adb2e6cfb45-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.800545 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" event={"ID":"78f0df6a-316c-4970-b612-3adb2e6cfb45","Type":"ContainerDied","Data":"854d47d4492f0360352fbfff121d08b2ca44a0f93f56cac3618921646816c359"} Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.800807 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="854d47d4492f0360352fbfff121d08b2ca44a0f93f56cac3618921646816c359" Dec 06 08:00:03 crc kubenswrapper[4945]: I1206 08:00:03.800634 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng" Dec 06 08:00:04 crc kubenswrapper[4945]: I1206 08:00:04.210194 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs"] Dec 06 08:00:04 crc kubenswrapper[4945]: I1206 08:00:04.215267 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416755-q9gfs"] Dec 06 08:00:04 crc kubenswrapper[4945]: I1206 08:00:04.960391 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="359894c8-e451-4b8c-a0ed-48902377d481" path="/var/lib/kubelet/pods/359894c8-e451-4b8c-a0ed-48902377d481/volumes" Dec 06 08:00:11 crc kubenswrapper[4945]: I1206 08:00:11.953461 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:00:11 crc kubenswrapper[4945]: E1206 08:00:11.954368 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:00:18 crc kubenswrapper[4945]: I1206 08:00:18.339345 4945 scope.go:117] "RemoveContainer" containerID="640e52dff2a20bc3c593474e6eb40ae052b77fc371b61a009179e281519d3245" Dec 06 08:00:23 crc kubenswrapper[4945]: I1206 08:00:23.953217 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:00:23 crc kubenswrapper[4945]: E1206 08:00:23.954194 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:00:37 crc kubenswrapper[4945]: I1206 08:00:37.953342 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:00:37 crc kubenswrapper[4945]: E1206 08:00:37.953986 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:00:49 crc kubenswrapper[4945]: I1206 08:00:49.954010 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:00:49 crc kubenswrapper[4945]: E1206 08:00:49.958031 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:01:01 crc kubenswrapper[4945]: I1206 08:01:01.954204 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:01:01 crc kubenswrapper[4945]: E1206 08:01:01.955120 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:01:13 crc kubenswrapper[4945]: I1206 08:01:13.954777 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:01:13 crc kubenswrapper[4945]: E1206 08:01:13.956526 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:01:28 crc kubenswrapper[4945]: I1206 08:01:28.953825 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:01:28 crc kubenswrapper[4945]: E1206 08:01:28.954679 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:01:42 crc kubenswrapper[4945]: I1206 08:01:42.953297 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:01:42 crc kubenswrapper[4945]: E1206 08:01:42.954022 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:01:55 crc kubenswrapper[4945]: I1206 08:01:55.953692 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:01:55 crc kubenswrapper[4945]: E1206 08:01:55.954422 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:02:03 crc kubenswrapper[4945]: I1206 08:02:03.743594 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hlhzm"] Dec 06 08:02:03 crc kubenswrapper[4945]: E1206 08:02:03.744425 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f0df6a-316c-4970-b612-3adb2e6cfb45" containerName="collect-profiles" Dec 06 08:02:03 crc kubenswrapper[4945]: I1206 08:02:03.744440 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f0df6a-316c-4970-b612-3adb2e6cfb45" containerName="collect-profiles" Dec 06 08:02:03 crc kubenswrapper[4945]: I1206 08:02:03.744572 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="78f0df6a-316c-4970-b612-3adb2e6cfb45" containerName="collect-profiles" Dec 06 08:02:03 crc kubenswrapper[4945]: I1206 08:02:03.745718 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:03 crc kubenswrapper[4945]: I1206 08:02:03.762997 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hlhzm"] Dec 06 08:02:03 crc kubenswrapper[4945]: I1206 08:02:03.924451 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-catalog-content\") pod \"community-operators-hlhzm\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:03 crc kubenswrapper[4945]: I1206 08:02:03.924518 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrqdq\" (UniqueName: \"kubernetes.io/projected/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-kube-api-access-nrqdq\") pod \"community-operators-hlhzm\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:03 crc kubenswrapper[4945]: I1206 08:02:03.924546 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-utilities\") pod \"community-operators-hlhzm\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:04 crc kubenswrapper[4945]: I1206 08:02:04.026106 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrqdq\" (UniqueName: \"kubernetes.io/projected/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-kube-api-access-nrqdq\") pod \"community-operators-hlhzm\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:04 crc kubenswrapper[4945]: I1206 08:02:04.026187 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-utilities\") pod \"community-operators-hlhzm\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:04 crc kubenswrapper[4945]: I1206 08:02:04.026320 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-catalog-content\") pod \"community-operators-hlhzm\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:04 crc kubenswrapper[4945]: I1206 08:02:04.026819 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-utilities\") pod \"community-operators-hlhzm\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:04 crc kubenswrapper[4945]: I1206 08:02:04.026886 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-catalog-content\") pod \"community-operators-hlhzm\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:04 crc kubenswrapper[4945]: I1206 08:02:04.046459 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrqdq\" (UniqueName: \"kubernetes.io/projected/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-kube-api-access-nrqdq\") pod \"community-operators-hlhzm\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:04 crc kubenswrapper[4945]: I1206 08:02:04.070646 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:04 crc kubenswrapper[4945]: I1206 08:02:04.541147 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hlhzm"] Dec 06 08:02:04 crc kubenswrapper[4945]: I1206 08:02:04.675952 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hlhzm" event={"ID":"b70b57ed-7942-4cdd-a05b-364ad88c5fa7","Type":"ContainerStarted","Data":"130a2984a02fa877d1c251479d3d44bc5215b6fe61df4ddeb91a1fddf942cad3"} Dec 06 08:02:05 crc kubenswrapper[4945]: I1206 08:02:05.686691 4945 generic.go:334] "Generic (PLEG): container finished" podID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerID="85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829" exitCode=0 Dec 06 08:02:05 crc kubenswrapper[4945]: I1206 08:02:05.686735 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hlhzm" event={"ID":"b70b57ed-7942-4cdd-a05b-364ad88c5fa7","Type":"ContainerDied","Data":"85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829"} Dec 06 08:02:07 crc kubenswrapper[4945]: I1206 08:02:07.703230 4945 generic.go:334] "Generic (PLEG): container finished" podID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerID="2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228" exitCode=0 Dec 06 08:02:07 crc kubenswrapper[4945]: I1206 08:02:07.703356 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hlhzm" event={"ID":"b70b57ed-7942-4cdd-a05b-364ad88c5fa7","Type":"ContainerDied","Data":"2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228"} Dec 06 08:02:08 crc kubenswrapper[4945]: I1206 08:02:08.714892 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hlhzm" event={"ID":"b70b57ed-7942-4cdd-a05b-364ad88c5fa7","Type":"ContainerStarted","Data":"f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96"} Dec 06 08:02:08 crc kubenswrapper[4945]: I1206 08:02:08.953312 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:02:09 crc kubenswrapper[4945]: I1206 08:02:09.724014 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"6bd4a6f6505690c4b9d7a9a3b7401b0aa12c4aa77c0b2a5ba56c5fa3f19605bb"} Dec 06 08:02:09 crc kubenswrapper[4945]: I1206 08:02:09.746381 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hlhzm" podStartSLOduration=4.1257219 podStartE2EDuration="6.746361647s" podCreationTimestamp="2025-12-06 08:02:03 +0000 UTC" firstStartedPulling="2025-12-06 08:02:05.688894964 +0000 UTC m=+4139.143756008" lastFinishedPulling="2025-12-06 08:02:08.309534711 +0000 UTC m=+4141.764395755" observedRunningTime="2025-12-06 08:02:09.740212333 +0000 UTC m=+4143.195073387" watchObservedRunningTime="2025-12-06 08:02:09.746361647 +0000 UTC m=+4143.201222721" Dec 06 08:02:14 crc kubenswrapper[4945]: I1206 08:02:14.071345 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:14 crc kubenswrapper[4945]: I1206 08:02:14.072466 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:14 crc kubenswrapper[4945]: I1206 08:02:14.113536 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:14 crc kubenswrapper[4945]: I1206 08:02:14.811832 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:14 crc kubenswrapper[4945]: I1206 08:02:14.864812 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hlhzm"] Dec 06 08:02:16 crc kubenswrapper[4945]: I1206 08:02:16.775644 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hlhzm" podUID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerName="registry-server" containerID="cri-o://f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96" gracePeriod=2 Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.364004 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.527360 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-catalog-content\") pod \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.527408 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-utilities\") pod \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.527454 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrqdq\" (UniqueName: \"kubernetes.io/projected/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-kube-api-access-nrqdq\") pod \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\" (UID: \"b70b57ed-7942-4cdd-a05b-364ad88c5fa7\") " Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.528897 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-utilities" (OuterVolumeSpecName: "utilities") pod "b70b57ed-7942-4cdd-a05b-364ad88c5fa7" (UID: "b70b57ed-7942-4cdd-a05b-364ad88c5fa7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.537630 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-kube-api-access-nrqdq" (OuterVolumeSpecName: "kube-api-access-nrqdq") pod "b70b57ed-7942-4cdd-a05b-364ad88c5fa7" (UID: "b70b57ed-7942-4cdd-a05b-364ad88c5fa7"). InnerVolumeSpecName "kube-api-access-nrqdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.591511 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b70b57ed-7942-4cdd-a05b-364ad88c5fa7" (UID: "b70b57ed-7942-4cdd-a05b-364ad88c5fa7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.629471 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.629508 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.629545 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrqdq\" (UniqueName: \"kubernetes.io/projected/b70b57ed-7942-4cdd-a05b-364ad88c5fa7-kube-api-access-nrqdq\") on node \"crc\" DevicePath \"\"" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.785927 4945 generic.go:334] "Generic (PLEG): container finished" podID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerID="f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96" exitCode=0 Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.785980 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hlhzm" event={"ID":"b70b57ed-7942-4cdd-a05b-364ad88c5fa7","Type":"ContainerDied","Data":"f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96"} Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.786072 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hlhzm" event={"ID":"b70b57ed-7942-4cdd-a05b-364ad88c5fa7","Type":"ContainerDied","Data":"130a2984a02fa877d1c251479d3d44bc5215b6fe61df4ddeb91a1fddf942cad3"} Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.786089 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hlhzm" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.786101 4945 scope.go:117] "RemoveContainer" containerID="f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.807579 4945 scope.go:117] "RemoveContainer" containerID="2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.832888 4945 scope.go:117] "RemoveContainer" containerID="85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.833040 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hlhzm"] Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.838811 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hlhzm"] Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.855104 4945 scope.go:117] "RemoveContainer" containerID="f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96" Dec 06 08:02:17 crc kubenswrapper[4945]: E1206 08:02:17.855740 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96\": container with ID starting with f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96 not found: ID does not exist" containerID="f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.855798 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96"} err="failed to get container status \"f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96\": rpc error: code = NotFound desc = could not find container \"f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96\": container with ID starting with f43e5aa688c0108e44525026c905a341a88910e3aedb623f0959a8482ec1dd96 not found: ID does not exist" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.855828 4945 scope.go:117] "RemoveContainer" containerID="2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228" Dec 06 08:02:17 crc kubenswrapper[4945]: E1206 08:02:17.856180 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228\": container with ID starting with 2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228 not found: ID does not exist" containerID="2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.856230 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228"} err="failed to get container status \"2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228\": rpc error: code = NotFound desc = could not find container \"2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228\": container with ID starting with 2d2cecc9b3c7ab7556fd966226a90c0dc8c22d141f406bf946593126e9f67228 not found: ID does not exist" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.856253 4945 scope.go:117] "RemoveContainer" containerID="85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829" Dec 06 08:02:17 crc kubenswrapper[4945]: E1206 08:02:17.856492 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829\": container with ID starting with 85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829 not found: ID does not exist" containerID="85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829" Dec 06 08:02:17 crc kubenswrapper[4945]: I1206 08:02:17.856524 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829"} err="failed to get container status \"85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829\": rpc error: code = NotFound desc = could not find container \"85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829\": container with ID starting with 85130a5df921d50b2a199dfec915c1af209114c7bbecc6b3ec1e3b1a8fadf829 not found: ID does not exist" Dec 06 08:02:18 crc kubenswrapper[4945]: I1206 08:02:18.963213 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" path="/var/lib/kubelet/pods/b70b57ed-7942-4cdd-a05b-364ad88c5fa7/volumes" Dec 06 08:03:57 crc kubenswrapper[4945]: I1206 08:03:57.996076 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wzmbn"] Dec 06 08:03:57 crc kubenswrapper[4945]: E1206 08:03:57.997370 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerName="extract-utilities" Dec 06 08:03:57 crc kubenswrapper[4945]: I1206 08:03:57.997393 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerName="extract-utilities" Dec 06 08:03:57 crc kubenswrapper[4945]: E1206 08:03:57.997421 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerName="extract-content" Dec 06 08:03:57 crc kubenswrapper[4945]: I1206 08:03:57.997433 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerName="extract-content" Dec 06 08:03:57 crc kubenswrapper[4945]: E1206 08:03:57.997451 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerName="registry-server" Dec 06 08:03:57 crc kubenswrapper[4945]: I1206 08:03:57.997462 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerName="registry-server" Dec 06 08:03:57 crc kubenswrapper[4945]: I1206 08:03:57.997716 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b70b57ed-7942-4cdd-a05b-364ad88c5fa7" containerName="registry-server" Dec 06 08:03:57 crc kubenswrapper[4945]: I1206 08:03:57.999331 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.021137 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wzmbn"] Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.081973 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c6q2\" (UniqueName: \"kubernetes.io/projected/e08bae76-4747-4ca5-9039-ccbea5dadad2-kube-api-access-7c6q2\") pod \"redhat-operators-wzmbn\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.082061 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-utilities\") pod \"redhat-operators-wzmbn\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.082091 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-catalog-content\") pod \"redhat-operators-wzmbn\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.183690 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c6q2\" (UniqueName: \"kubernetes.io/projected/e08bae76-4747-4ca5-9039-ccbea5dadad2-kube-api-access-7c6q2\") pod \"redhat-operators-wzmbn\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.183825 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-utilities\") pod \"redhat-operators-wzmbn\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.183861 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-catalog-content\") pod \"redhat-operators-wzmbn\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.184348 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-utilities\") pod \"redhat-operators-wzmbn\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.184590 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-catalog-content\") pod \"redhat-operators-wzmbn\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.223121 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c6q2\" (UniqueName: \"kubernetes.io/projected/e08bae76-4747-4ca5-9039-ccbea5dadad2-kube-api-access-7c6q2\") pod \"redhat-operators-wzmbn\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.319045 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:03:58 crc kubenswrapper[4945]: I1206 08:03:58.761028 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wzmbn"] Dec 06 08:03:59 crc kubenswrapper[4945]: I1206 08:03:59.688773 4945 generic.go:334] "Generic (PLEG): container finished" podID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerID="1a32df848d8ca1a04f0453cd4557a40292f5ba7f482e6b74bc9ef26e3b812a4f" exitCode=0 Dec 06 08:03:59 crc kubenswrapper[4945]: I1206 08:03:59.688848 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzmbn" event={"ID":"e08bae76-4747-4ca5-9039-ccbea5dadad2","Type":"ContainerDied","Data":"1a32df848d8ca1a04f0453cd4557a40292f5ba7f482e6b74bc9ef26e3b812a4f"} Dec 06 08:03:59 crc kubenswrapper[4945]: I1206 08:03:59.688873 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzmbn" event={"ID":"e08bae76-4747-4ca5-9039-ccbea5dadad2","Type":"ContainerStarted","Data":"c6bbc215bda20167dfd61500100295cb0c122d73e9011209f07de73b39359a47"} Dec 06 08:03:59 crc kubenswrapper[4945]: I1206 08:03:59.691520 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 08:04:01 crc kubenswrapper[4945]: I1206 08:04:01.706899 4945 generic.go:334] "Generic (PLEG): container finished" podID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerID="c05d7748c00ea539e5020fa5db7c644c37b0cb295d85db8c801ea781ecc2e3a8" exitCode=0 Dec 06 08:04:01 crc kubenswrapper[4945]: I1206 08:04:01.707345 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzmbn" event={"ID":"e08bae76-4747-4ca5-9039-ccbea5dadad2","Type":"ContainerDied","Data":"c05d7748c00ea539e5020fa5db7c644c37b0cb295d85db8c801ea781ecc2e3a8"} Dec 06 08:04:02 crc kubenswrapper[4945]: I1206 08:04:02.715218 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzmbn" event={"ID":"e08bae76-4747-4ca5-9039-ccbea5dadad2","Type":"ContainerStarted","Data":"69cd9f486142de83943e5cab00c770136a03b03b54871489c7de8bd6ef28b3dd"} Dec 06 08:04:02 crc kubenswrapper[4945]: I1206 08:04:02.751380 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wzmbn" podStartSLOduration=3.281683659 podStartE2EDuration="5.751356021s" podCreationTimestamp="2025-12-06 08:03:57 +0000 UTC" firstStartedPulling="2025-12-06 08:03:59.691307838 +0000 UTC m=+4253.146168882" lastFinishedPulling="2025-12-06 08:04:02.1609802 +0000 UTC m=+4255.615841244" observedRunningTime="2025-12-06 08:04:02.747305273 +0000 UTC m=+4256.202166317" watchObservedRunningTime="2025-12-06 08:04:02.751356021 +0000 UTC m=+4256.206217075" Dec 06 08:04:08 crc kubenswrapper[4945]: I1206 08:04:08.319476 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:04:08 crc kubenswrapper[4945]: I1206 08:04:08.320026 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:04:08 crc kubenswrapper[4945]: I1206 08:04:08.370854 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:04:08 crc kubenswrapper[4945]: I1206 08:04:08.811109 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:04:08 crc kubenswrapper[4945]: I1206 08:04:08.865371 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wzmbn"] Dec 06 08:04:10 crc kubenswrapper[4945]: I1206 08:04:10.772649 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wzmbn" podUID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerName="registry-server" containerID="cri-o://69cd9f486142de83943e5cab00c770136a03b03b54871489c7de8bd6ef28b3dd" gracePeriod=2 Dec 06 08:04:13 crc kubenswrapper[4945]: I1206 08:04:13.805835 4945 generic.go:334] "Generic (PLEG): container finished" podID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerID="69cd9f486142de83943e5cab00c770136a03b03b54871489c7de8bd6ef28b3dd" exitCode=0 Dec 06 08:04:13 crc kubenswrapper[4945]: I1206 08:04:13.805925 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzmbn" event={"ID":"e08bae76-4747-4ca5-9039-ccbea5dadad2","Type":"ContainerDied","Data":"69cd9f486142de83943e5cab00c770136a03b03b54871489c7de8bd6ef28b3dd"} Dec 06 08:04:13 crc kubenswrapper[4945]: I1206 08:04:13.888169 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.016904 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-catalog-content\") pod \"e08bae76-4747-4ca5-9039-ccbea5dadad2\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.017024 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c6q2\" (UniqueName: \"kubernetes.io/projected/e08bae76-4747-4ca5-9039-ccbea5dadad2-kube-api-access-7c6q2\") pod \"e08bae76-4747-4ca5-9039-ccbea5dadad2\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.017080 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-utilities\") pod \"e08bae76-4747-4ca5-9039-ccbea5dadad2\" (UID: \"e08bae76-4747-4ca5-9039-ccbea5dadad2\") " Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.018192 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-utilities" (OuterVolumeSpecName: "utilities") pod "e08bae76-4747-4ca5-9039-ccbea5dadad2" (UID: "e08bae76-4747-4ca5-9039-ccbea5dadad2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.030667 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e08bae76-4747-4ca5-9039-ccbea5dadad2-kube-api-access-7c6q2" (OuterVolumeSpecName: "kube-api-access-7c6q2") pod "e08bae76-4747-4ca5-9039-ccbea5dadad2" (UID: "e08bae76-4747-4ca5-9039-ccbea5dadad2"). InnerVolumeSpecName "kube-api-access-7c6q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.118955 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c6q2\" (UniqueName: \"kubernetes.io/projected/e08bae76-4747-4ca5-9039-ccbea5dadad2-kube-api-access-7c6q2\") on node \"crc\" DevicePath \"\"" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.118985 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.125996 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e08bae76-4747-4ca5-9039-ccbea5dadad2" (UID: "e08bae76-4747-4ca5-9039-ccbea5dadad2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.221192 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e08bae76-4747-4ca5-9039-ccbea5dadad2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.816933 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzmbn" event={"ID":"e08bae76-4747-4ca5-9039-ccbea5dadad2","Type":"ContainerDied","Data":"c6bbc215bda20167dfd61500100295cb0c122d73e9011209f07de73b39359a47"} Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.817590 4945 scope.go:117] "RemoveContainer" containerID="69cd9f486142de83943e5cab00c770136a03b03b54871489c7de8bd6ef28b3dd" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.817073 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wzmbn" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.840089 4945 scope.go:117] "RemoveContainer" containerID="c05d7748c00ea539e5020fa5db7c644c37b0cb295d85db8c801ea781ecc2e3a8" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.872570 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wzmbn"] Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.883672 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wzmbn"] Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.893176 4945 scope.go:117] "RemoveContainer" containerID="1a32df848d8ca1a04f0453cd4557a40292f5ba7f482e6b74bc9ef26e3b812a4f" Dec 06 08:04:14 crc kubenswrapper[4945]: I1206 08:04:14.960980 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e08bae76-4747-4ca5-9039-ccbea5dadad2" path="/var/lib/kubelet/pods/e08bae76-4747-4ca5-9039-ccbea5dadad2/volumes" Dec 06 08:04:38 crc kubenswrapper[4945]: I1206 08:04:38.795788 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:04:38 crc kubenswrapper[4945]: I1206 08:04:38.796514 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:05:08 crc kubenswrapper[4945]: I1206 08:05:08.795646 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:05:08 crc kubenswrapper[4945]: I1206 08:05:08.796354 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.528811 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-85qdr"] Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.534617 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-85qdr"] Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.704028 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-96k2q"] Dec 06 08:05:37 crc kubenswrapper[4945]: E1206 08:05:37.704331 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerName="registry-server" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.704346 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerName="registry-server" Dec 06 08:05:37 crc kubenswrapper[4945]: E1206 08:05:37.704356 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerName="extract-content" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.704363 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerName="extract-content" Dec 06 08:05:37 crc kubenswrapper[4945]: E1206 08:05:37.704376 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerName="extract-utilities" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.704382 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerName="extract-utilities" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.704522 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e08bae76-4747-4ca5-9039-ccbea5dadad2" containerName="registry-server" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.705430 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.707366 4945 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-mqlkm" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.707634 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.710082 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.710382 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.718087 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-96k2q"] Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.838707 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-node-mnt\") pod \"crc-storage-crc-96k2q\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.838775 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-448mt\" (UniqueName: \"kubernetes.io/projected/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-kube-api-access-448mt\") pod \"crc-storage-crc-96k2q\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.839068 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-crc-storage\") pod \"crc-storage-crc-96k2q\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.940788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-crc-storage\") pod \"crc-storage-crc-96k2q\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.940852 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-node-mnt\") pod \"crc-storage-crc-96k2q\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.940895 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-448mt\" (UniqueName: \"kubernetes.io/projected/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-kube-api-access-448mt\") pod \"crc-storage-crc-96k2q\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.941162 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-node-mnt\") pod \"crc-storage-crc-96k2q\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.941699 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-crc-storage\") pod \"crc-storage-crc-96k2q\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:37 crc kubenswrapper[4945]: I1206 08:05:37.960208 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-448mt\" (UniqueName: \"kubernetes.io/projected/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-kube-api-access-448mt\") pod \"crc-storage-crc-96k2q\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:38 crc kubenswrapper[4945]: I1206 08:05:38.023954 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:38 crc kubenswrapper[4945]: I1206 08:05:38.450312 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-96k2q"] Dec 06 08:05:38 crc kubenswrapper[4945]: I1206 08:05:38.489680 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-96k2q" event={"ID":"78ad4304-2f66-48b3-9dce-5e9b6e4e4347","Type":"ContainerStarted","Data":"db2d905aef494dde2d2580dc84a1e48826043171ca24d70843385bb0b2b87231"} Dec 06 08:05:38 crc kubenswrapper[4945]: I1206 08:05:38.796050 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:05:38 crc kubenswrapper[4945]: I1206 08:05:38.796975 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:05:38 crc kubenswrapper[4945]: I1206 08:05:38.797115 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 08:05:38 crc kubenswrapper[4945]: I1206 08:05:38.797712 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6bd4a6f6505690c4b9d7a9a3b7401b0aa12c4aa77c0b2a5ba56c5fa3f19605bb"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 08:05:38 crc kubenswrapper[4945]: I1206 08:05:38.797849 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://6bd4a6f6505690c4b9d7a9a3b7401b0aa12c4aa77c0b2a5ba56c5fa3f19605bb" gracePeriod=600 Dec 06 08:05:38 crc kubenswrapper[4945]: I1206 08:05:38.962644 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42d31853-9fa8-4d5d-85fc-db2179129bfe" path="/var/lib/kubelet/pods/42d31853-9fa8-4d5d-85fc-db2179129bfe/volumes" Dec 06 08:05:39 crc kubenswrapper[4945]: I1206 08:05:39.498008 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-96k2q" event={"ID":"78ad4304-2f66-48b3-9dce-5e9b6e4e4347","Type":"ContainerStarted","Data":"6231ab2b9e48d0041945fc8631173f9a79801396f6f21724b5f1980ca7f7dbce"} Dec 06 08:05:39 crc kubenswrapper[4945]: I1206 08:05:39.501317 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="6bd4a6f6505690c4b9d7a9a3b7401b0aa12c4aa77c0b2a5ba56c5fa3f19605bb" exitCode=0 Dec 06 08:05:39 crc kubenswrapper[4945]: I1206 08:05:39.501358 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"6bd4a6f6505690c4b9d7a9a3b7401b0aa12c4aa77c0b2a5ba56c5fa3f19605bb"} Dec 06 08:05:39 crc kubenswrapper[4945]: I1206 08:05:39.501384 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4"} Dec 06 08:05:39 crc kubenswrapper[4945]: I1206 08:05:39.501400 4945 scope.go:117] "RemoveContainer" containerID="cff99fc22330f58cee154f4557771f1f83067cc4da13943e8544b6eedfebc7b8" Dec 06 08:05:39 crc kubenswrapper[4945]: I1206 08:05:39.518715 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-96k2q" podStartSLOduration=1.902821403 podStartE2EDuration="2.518698648s" podCreationTimestamp="2025-12-06 08:05:37 +0000 UTC" firstStartedPulling="2025-12-06 08:05:38.456950965 +0000 UTC m=+4351.911812009" lastFinishedPulling="2025-12-06 08:05:39.07282821 +0000 UTC m=+4352.527689254" observedRunningTime="2025-12-06 08:05:39.518616666 +0000 UTC m=+4352.973477700" watchObservedRunningTime="2025-12-06 08:05:39.518698648 +0000 UTC m=+4352.973559692" Dec 06 08:05:40 crc kubenswrapper[4945]: I1206 08:05:40.510139 4945 generic.go:334] "Generic (PLEG): container finished" podID="78ad4304-2f66-48b3-9dce-5e9b6e4e4347" containerID="6231ab2b9e48d0041945fc8631173f9a79801396f6f21724b5f1980ca7f7dbce" exitCode=0 Dec 06 08:05:40 crc kubenswrapper[4945]: I1206 08:05:40.510404 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-96k2q" event={"ID":"78ad4304-2f66-48b3-9dce-5e9b6e4e4347","Type":"ContainerDied","Data":"6231ab2b9e48d0041945fc8631173f9a79801396f6f21724b5f1980ca7f7dbce"} Dec 06 08:05:41 crc kubenswrapper[4945]: I1206 08:05:41.767637 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:41 crc kubenswrapper[4945]: I1206 08:05:41.913937 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-node-mnt\") pod \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " Dec 06 08:05:41 crc kubenswrapper[4945]: I1206 08:05:41.914072 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "78ad4304-2f66-48b3-9dce-5e9b6e4e4347" (UID: "78ad4304-2f66-48b3-9dce-5e9b6e4e4347"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 08:05:41 crc kubenswrapper[4945]: I1206 08:05:41.914095 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-crc-storage\") pod \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " Dec 06 08:05:41 crc kubenswrapper[4945]: I1206 08:05:41.914136 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-448mt\" (UniqueName: \"kubernetes.io/projected/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-kube-api-access-448mt\") pod \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\" (UID: \"78ad4304-2f66-48b3-9dce-5e9b6e4e4347\") " Dec 06 08:05:41 crc kubenswrapper[4945]: I1206 08:05:41.914568 4945 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 06 08:05:41 crc kubenswrapper[4945]: I1206 08:05:41.918793 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-kube-api-access-448mt" (OuterVolumeSpecName: "kube-api-access-448mt") pod "78ad4304-2f66-48b3-9dce-5e9b6e4e4347" (UID: "78ad4304-2f66-48b3-9dce-5e9b6e4e4347"). InnerVolumeSpecName "kube-api-access-448mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:05:41 crc kubenswrapper[4945]: I1206 08:05:41.930499 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "78ad4304-2f66-48b3-9dce-5e9b6e4e4347" (UID: "78ad4304-2f66-48b3-9dce-5e9b6e4e4347"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:05:42 crc kubenswrapper[4945]: I1206 08:05:42.015497 4945 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 06 08:05:42 crc kubenswrapper[4945]: I1206 08:05:42.015548 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-448mt\" (UniqueName: \"kubernetes.io/projected/78ad4304-2f66-48b3-9dce-5e9b6e4e4347-kube-api-access-448mt\") on node \"crc\" DevicePath \"\"" Dec 06 08:05:42 crc kubenswrapper[4945]: I1206 08:05:42.534121 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-96k2q" event={"ID":"78ad4304-2f66-48b3-9dce-5e9b6e4e4347","Type":"ContainerDied","Data":"db2d905aef494dde2d2580dc84a1e48826043171ca24d70843385bb0b2b87231"} Dec 06 08:05:42 crc kubenswrapper[4945]: I1206 08:05:42.534190 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db2d905aef494dde2d2580dc84a1e48826043171ca24d70843385bb0b2b87231" Dec 06 08:05:42 crc kubenswrapper[4945]: I1206 08:05:42.534276 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-96k2q" Dec 06 08:05:43 crc kubenswrapper[4945]: I1206 08:05:43.996085 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-96k2q"] Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.001907 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-96k2q"] Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.176556 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-rkbgw"] Dec 06 08:05:44 crc kubenswrapper[4945]: E1206 08:05:44.176889 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78ad4304-2f66-48b3-9dce-5e9b6e4e4347" containerName="storage" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.176906 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="78ad4304-2f66-48b3-9dce-5e9b6e4e4347" containerName="storage" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.177039 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="78ad4304-2f66-48b3-9dce-5e9b6e4e4347" containerName="storage" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.177549 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.180091 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.181049 4945 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-mqlkm" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.181110 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.181165 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.191686 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-rkbgw"] Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.347706 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz9z7\" (UniqueName: \"kubernetes.io/projected/b6cca087-b96b-499a-a3a4-312218cc61c2-kube-api-access-fz9z7\") pod \"crc-storage-crc-rkbgw\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.348048 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b6cca087-b96b-499a-a3a4-312218cc61c2-node-mnt\") pod \"crc-storage-crc-rkbgw\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.348092 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b6cca087-b96b-499a-a3a4-312218cc61c2-crc-storage\") pod \"crc-storage-crc-rkbgw\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.448920 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b6cca087-b96b-499a-a3a4-312218cc61c2-crc-storage\") pod \"crc-storage-crc-rkbgw\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.448990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz9z7\" (UniqueName: \"kubernetes.io/projected/b6cca087-b96b-499a-a3a4-312218cc61c2-kube-api-access-fz9z7\") pod \"crc-storage-crc-rkbgw\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.449062 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b6cca087-b96b-499a-a3a4-312218cc61c2-node-mnt\") pod \"crc-storage-crc-rkbgw\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.449417 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b6cca087-b96b-499a-a3a4-312218cc61c2-node-mnt\") pod \"crc-storage-crc-rkbgw\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.454694 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b6cca087-b96b-499a-a3a4-312218cc61c2-crc-storage\") pod \"crc-storage-crc-rkbgw\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.470255 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz9z7\" (UniqueName: \"kubernetes.io/projected/b6cca087-b96b-499a-a3a4-312218cc61c2-kube-api-access-fz9z7\") pod \"crc-storage-crc-rkbgw\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.513738 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.918097 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-rkbgw"] Dec 06 08:05:44 crc kubenswrapper[4945]: I1206 08:05:44.963338 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78ad4304-2f66-48b3-9dce-5e9b6e4e4347" path="/var/lib/kubelet/pods/78ad4304-2f66-48b3-9dce-5e9b6e4e4347/volumes" Dec 06 08:05:45 crc kubenswrapper[4945]: I1206 08:05:45.554326 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-rkbgw" event={"ID":"b6cca087-b96b-499a-a3a4-312218cc61c2","Type":"ContainerStarted","Data":"8c312bef3fc5ef8b98e7e25fc633be68793a6b4352f7302716d634480b0ec2ed"} Dec 06 08:05:47 crc kubenswrapper[4945]: I1206 08:05:47.568764 4945 generic.go:334] "Generic (PLEG): container finished" podID="b6cca087-b96b-499a-a3a4-312218cc61c2" containerID="bc54d4e254672c3a497c1c02bccb3b120104748e4f3629371c15569315dbfebf" exitCode=0 Dec 06 08:05:47 crc kubenswrapper[4945]: I1206 08:05:47.568914 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-rkbgw" event={"ID":"b6cca087-b96b-499a-a3a4-312218cc61c2","Type":"ContainerDied","Data":"bc54d4e254672c3a497c1c02bccb3b120104748e4f3629371c15569315dbfebf"} Dec 06 08:05:48 crc kubenswrapper[4945]: I1206 08:05:48.822249 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:05:48 crc kubenswrapper[4945]: I1206 08:05:48.907705 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz9z7\" (UniqueName: \"kubernetes.io/projected/b6cca087-b96b-499a-a3a4-312218cc61c2-kube-api-access-fz9z7\") pod \"b6cca087-b96b-499a-a3a4-312218cc61c2\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " Dec 06 08:05:48 crc kubenswrapper[4945]: I1206 08:05:48.907859 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b6cca087-b96b-499a-a3a4-312218cc61c2-node-mnt\") pod \"b6cca087-b96b-499a-a3a4-312218cc61c2\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " Dec 06 08:05:48 crc kubenswrapper[4945]: I1206 08:05:48.907924 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b6cca087-b96b-499a-a3a4-312218cc61c2-crc-storage\") pod \"b6cca087-b96b-499a-a3a4-312218cc61c2\" (UID: \"b6cca087-b96b-499a-a3a4-312218cc61c2\") " Dec 06 08:05:48 crc kubenswrapper[4945]: I1206 08:05:48.907987 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6cca087-b96b-499a-a3a4-312218cc61c2-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "b6cca087-b96b-499a-a3a4-312218cc61c2" (UID: "b6cca087-b96b-499a-a3a4-312218cc61c2"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 08:05:48 crc kubenswrapper[4945]: I1206 08:05:48.908217 4945 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b6cca087-b96b-499a-a3a4-312218cc61c2-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 06 08:05:48 crc kubenswrapper[4945]: I1206 08:05:48.919599 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cca087-b96b-499a-a3a4-312218cc61c2-kube-api-access-fz9z7" (OuterVolumeSpecName: "kube-api-access-fz9z7") pod "b6cca087-b96b-499a-a3a4-312218cc61c2" (UID: "b6cca087-b96b-499a-a3a4-312218cc61c2"). InnerVolumeSpecName "kube-api-access-fz9z7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:05:48 crc kubenswrapper[4945]: I1206 08:05:48.925676 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cca087-b96b-499a-a3a4-312218cc61c2-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "b6cca087-b96b-499a-a3a4-312218cc61c2" (UID: "b6cca087-b96b-499a-a3a4-312218cc61c2"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:05:49 crc kubenswrapper[4945]: I1206 08:05:49.009258 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz9z7\" (UniqueName: \"kubernetes.io/projected/b6cca087-b96b-499a-a3a4-312218cc61c2-kube-api-access-fz9z7\") on node \"crc\" DevicePath \"\"" Dec 06 08:05:49 crc kubenswrapper[4945]: I1206 08:05:49.009317 4945 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b6cca087-b96b-499a-a3a4-312218cc61c2-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 06 08:05:49 crc kubenswrapper[4945]: I1206 08:05:49.584580 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-rkbgw" event={"ID":"b6cca087-b96b-499a-a3a4-312218cc61c2","Type":"ContainerDied","Data":"8c312bef3fc5ef8b98e7e25fc633be68793a6b4352f7302716d634480b0ec2ed"} Dec 06 08:05:49 crc kubenswrapper[4945]: I1206 08:05:49.584616 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c312bef3fc5ef8b98e7e25fc633be68793a6b4352f7302716d634480b0ec2ed" Dec 06 08:05:49 crc kubenswrapper[4945]: I1206 08:05:49.584680 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-rkbgw" Dec 06 08:06:18 crc kubenswrapper[4945]: I1206 08:06:18.482650 4945 scope.go:117] "RemoveContainer" containerID="d1a103b3b7982b4586c200a020c46be4223b800dabf858bae432a976d96a0efd" Dec 06 08:08:04 crc kubenswrapper[4945]: I1206 08:08:04.923275 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5cjnp"] Dec 06 08:08:04 crc kubenswrapper[4945]: E1206 08:08:04.924389 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6cca087-b96b-499a-a3a4-312218cc61c2" containerName="storage" Dec 06 08:08:04 crc kubenswrapper[4945]: I1206 08:08:04.924412 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6cca087-b96b-499a-a3a4-312218cc61c2" containerName="storage" Dec 06 08:08:04 crc kubenswrapper[4945]: I1206 08:08:04.924661 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6cca087-b96b-499a-a3a4-312218cc61c2" containerName="storage" Dec 06 08:08:04 crc kubenswrapper[4945]: I1206 08:08:04.929320 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:04 crc kubenswrapper[4945]: I1206 08:08:04.936043 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5cjnp"] Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.041986 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-catalog-content\") pod \"redhat-marketplace-5cjnp\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.042063 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-utilities\") pod \"redhat-marketplace-5cjnp\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.042111 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdb69\" (UniqueName: \"kubernetes.io/projected/8d7c0122-19ff-4289-89b4-08a32e3ada60-kube-api-access-bdb69\") pod \"redhat-marketplace-5cjnp\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.143077 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdb69\" (UniqueName: \"kubernetes.io/projected/8d7c0122-19ff-4289-89b4-08a32e3ada60-kube-api-access-bdb69\") pod \"redhat-marketplace-5cjnp\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.143203 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-catalog-content\") pod \"redhat-marketplace-5cjnp\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.143273 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-utilities\") pod \"redhat-marketplace-5cjnp\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.144298 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-utilities\") pod \"redhat-marketplace-5cjnp\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.144487 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-catalog-content\") pod \"redhat-marketplace-5cjnp\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.167131 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdb69\" (UniqueName: \"kubernetes.io/projected/8d7c0122-19ff-4289-89b4-08a32e3ada60-kube-api-access-bdb69\") pod \"redhat-marketplace-5cjnp\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.264555 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:05 crc kubenswrapper[4945]: I1206 08:08:05.742121 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5cjnp"] Dec 06 08:08:06 crc kubenswrapper[4945]: I1206 08:08:06.632564 4945 generic.go:334] "Generic (PLEG): container finished" podID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerID="0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288" exitCode=0 Dec 06 08:08:06 crc kubenswrapper[4945]: I1206 08:08:06.633384 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cjnp" event={"ID":"8d7c0122-19ff-4289-89b4-08a32e3ada60","Type":"ContainerDied","Data":"0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288"} Dec 06 08:08:06 crc kubenswrapper[4945]: I1206 08:08:06.633439 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cjnp" event={"ID":"8d7c0122-19ff-4289-89b4-08a32e3ada60","Type":"ContainerStarted","Data":"aa80b85b88e3d59df4e87dba99c03d1be57b931d637ba5a8305b61491eb5db25"} Dec 06 08:08:08 crc kubenswrapper[4945]: I1206 08:08:08.648528 4945 generic.go:334] "Generic (PLEG): container finished" podID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerID="792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5" exitCode=0 Dec 06 08:08:08 crc kubenswrapper[4945]: I1206 08:08:08.648624 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cjnp" event={"ID":"8d7c0122-19ff-4289-89b4-08a32e3ada60","Type":"ContainerDied","Data":"792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5"} Dec 06 08:08:08 crc kubenswrapper[4945]: I1206 08:08:08.796225 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:08:08 crc kubenswrapper[4945]: I1206 08:08:08.796631 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:08:09 crc kubenswrapper[4945]: I1206 08:08:09.658402 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cjnp" event={"ID":"8d7c0122-19ff-4289-89b4-08a32e3ada60","Type":"ContainerStarted","Data":"f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac"} Dec 06 08:08:09 crc kubenswrapper[4945]: I1206 08:08:09.677244 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5cjnp" podStartSLOduration=3.268972128 podStartE2EDuration="5.677225557s" podCreationTimestamp="2025-12-06 08:08:04 +0000 UTC" firstStartedPulling="2025-12-06 08:08:06.635070532 +0000 UTC m=+4500.089931576" lastFinishedPulling="2025-12-06 08:08:09.043323961 +0000 UTC m=+4502.498185005" observedRunningTime="2025-12-06 08:08:09.676055866 +0000 UTC m=+4503.130916930" watchObservedRunningTime="2025-12-06 08:08:09.677225557 +0000 UTC m=+4503.132086601" Dec 06 08:08:15 crc kubenswrapper[4945]: I1206 08:08:15.265975 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:15 crc kubenswrapper[4945]: I1206 08:08:15.266742 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:15 crc kubenswrapper[4945]: I1206 08:08:15.330622 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:15 crc kubenswrapper[4945]: I1206 08:08:15.751442 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:15 crc kubenswrapper[4945]: I1206 08:08:15.805554 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5cjnp"] Dec 06 08:08:17 crc kubenswrapper[4945]: I1206 08:08:17.724096 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5cjnp" podUID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerName="registry-server" containerID="cri-o://f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac" gracePeriod=2 Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.496006 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.531453 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-utilities\") pod \"8d7c0122-19ff-4289-89b4-08a32e3ada60\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.531506 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-catalog-content\") pod \"8d7c0122-19ff-4289-89b4-08a32e3ada60\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.531553 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdb69\" (UniqueName: \"kubernetes.io/projected/8d7c0122-19ff-4289-89b4-08a32e3ada60-kube-api-access-bdb69\") pod \"8d7c0122-19ff-4289-89b4-08a32e3ada60\" (UID: \"8d7c0122-19ff-4289-89b4-08a32e3ada60\") " Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.532560 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-utilities" (OuterVolumeSpecName: "utilities") pod "8d7c0122-19ff-4289-89b4-08a32e3ada60" (UID: "8d7c0122-19ff-4289-89b4-08a32e3ada60"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.540513 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7c0122-19ff-4289-89b4-08a32e3ada60-kube-api-access-bdb69" (OuterVolumeSpecName: "kube-api-access-bdb69") pod "8d7c0122-19ff-4289-89b4-08a32e3ada60" (UID: "8d7c0122-19ff-4289-89b4-08a32e3ada60"). InnerVolumeSpecName "kube-api-access-bdb69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.552898 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d7c0122-19ff-4289-89b4-08a32e3ada60" (UID: "8d7c0122-19ff-4289-89b4-08a32e3ada60"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.633480 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.633536 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d7c0122-19ff-4289-89b4-08a32e3ada60-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.633549 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdb69\" (UniqueName: \"kubernetes.io/projected/8d7c0122-19ff-4289-89b4-08a32e3ada60-kube-api-access-bdb69\") on node \"crc\" DevicePath \"\"" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.732301 4945 generic.go:334] "Generic (PLEG): container finished" podID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerID="f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac" exitCode=0 Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.732364 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cjnp" event={"ID":"8d7c0122-19ff-4289-89b4-08a32e3ada60","Type":"ContainerDied","Data":"f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac"} Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.732392 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cjnp" event={"ID":"8d7c0122-19ff-4289-89b4-08a32e3ada60","Type":"ContainerDied","Data":"aa80b85b88e3d59df4e87dba99c03d1be57b931d637ba5a8305b61491eb5db25"} Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.732509 4945 scope.go:117] "RemoveContainer" containerID="f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.733248 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5cjnp" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.748877 4945 scope.go:117] "RemoveContainer" containerID="792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.765938 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5cjnp"] Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.772868 4945 scope.go:117] "RemoveContainer" containerID="0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.773061 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5cjnp"] Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.796250 4945 scope.go:117] "RemoveContainer" containerID="f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac" Dec 06 08:08:18 crc kubenswrapper[4945]: E1206 08:08:18.796831 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac\": container with ID starting with f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac not found: ID does not exist" containerID="f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.796861 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac"} err="failed to get container status \"f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac\": rpc error: code = NotFound desc = could not find container \"f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac\": container with ID starting with f7769d29ffd512366d60f27d43a63ec43cb9f6d55b3522138357445fdef4a5ac not found: ID does not exist" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.796886 4945 scope.go:117] "RemoveContainer" containerID="792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5" Dec 06 08:08:18 crc kubenswrapper[4945]: E1206 08:08:18.797565 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5\": container with ID starting with 792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5 not found: ID does not exist" containerID="792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.797584 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5"} err="failed to get container status \"792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5\": rpc error: code = NotFound desc = could not find container \"792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5\": container with ID starting with 792d3abd4b515216398385b8578cee6d616ed0df495bda442357f60d300855b5 not found: ID does not exist" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.797596 4945 scope.go:117] "RemoveContainer" containerID="0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288" Dec 06 08:08:18 crc kubenswrapper[4945]: E1206 08:08:18.797962 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288\": container with ID starting with 0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288 not found: ID does not exist" containerID="0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.797999 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288"} err="failed to get container status \"0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288\": rpc error: code = NotFound desc = could not find container \"0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288\": container with ID starting with 0238a80c55033461bb5624c9917fcfede1d14fe6b18141413d4741329592d288 not found: ID does not exist" Dec 06 08:08:18 crc kubenswrapper[4945]: I1206 08:08:18.963590 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d7c0122-19ff-4289-89b4-08a32e3ada60" path="/var/lib/kubelet/pods/8d7c0122-19ff-4289-89b4-08a32e3ada60/volumes" Dec 06 08:08:38 crc kubenswrapper[4945]: I1206 08:08:38.795819 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:08:38 crc kubenswrapper[4945]: I1206 08:08:38.796482 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.511741 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gp5hc"] Dec 06 08:08:41 crc kubenswrapper[4945]: E1206 08:08:41.512431 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerName="extract-utilities" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.512446 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerName="extract-utilities" Dec 06 08:08:41 crc kubenswrapper[4945]: E1206 08:08:41.512453 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerName="registry-server" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.512458 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerName="registry-server" Dec 06 08:08:41 crc kubenswrapper[4945]: E1206 08:08:41.512470 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerName="extract-content" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.512479 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerName="extract-content" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.512624 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7c0122-19ff-4289-89b4-08a32e3ada60" containerName="registry-server" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.513584 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.530402 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gp5hc"] Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.653462 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-catalog-content\") pod \"certified-operators-gp5hc\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.654121 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-utilities\") pod \"certified-operators-gp5hc\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.654341 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tzb9\" (UniqueName: \"kubernetes.io/projected/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-kube-api-access-9tzb9\") pod \"certified-operators-gp5hc\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.755927 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tzb9\" (UniqueName: \"kubernetes.io/projected/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-kube-api-access-9tzb9\") pod \"certified-operators-gp5hc\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.756001 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-catalog-content\") pod \"certified-operators-gp5hc\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.756091 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-utilities\") pod \"certified-operators-gp5hc\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.756687 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-utilities\") pod \"certified-operators-gp5hc\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.756811 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-catalog-content\") pod \"certified-operators-gp5hc\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.783402 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tzb9\" (UniqueName: \"kubernetes.io/projected/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-kube-api-access-9tzb9\") pod \"certified-operators-gp5hc\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:41 crc kubenswrapper[4945]: I1206 08:08:41.831619 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:42 crc kubenswrapper[4945]: I1206 08:08:42.031606 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gp5hc"] Dec 06 08:08:42 crc kubenswrapper[4945]: I1206 08:08:42.899266 4945 generic.go:334] "Generic (PLEG): container finished" podID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerID="0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034" exitCode=0 Dec 06 08:08:42 crc kubenswrapper[4945]: I1206 08:08:42.899364 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gp5hc" event={"ID":"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e","Type":"ContainerDied","Data":"0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034"} Dec 06 08:08:42 crc kubenswrapper[4945]: I1206 08:08:42.899602 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gp5hc" event={"ID":"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e","Type":"ContainerStarted","Data":"3bd557615f21225a5c072f6fab41758608a61c0cec2a40697bfe1e8904855f54"} Dec 06 08:08:43 crc kubenswrapper[4945]: I1206 08:08:43.906111 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gp5hc" event={"ID":"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e","Type":"ContainerStarted","Data":"6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb"} Dec 06 08:08:44 crc kubenswrapper[4945]: I1206 08:08:44.918022 4945 generic.go:334] "Generic (PLEG): container finished" podID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerID="6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb" exitCode=0 Dec 06 08:08:44 crc kubenswrapper[4945]: I1206 08:08:44.918092 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gp5hc" event={"ID":"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e","Type":"ContainerDied","Data":"6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb"} Dec 06 08:08:45 crc kubenswrapper[4945]: I1206 08:08:45.926762 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gp5hc" event={"ID":"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e","Type":"ContainerStarted","Data":"914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89"} Dec 06 08:08:45 crc kubenswrapper[4945]: I1206 08:08:45.951378 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gp5hc" podStartSLOduration=2.551690737 podStartE2EDuration="4.951358166s" podCreationTimestamp="2025-12-06 08:08:41 +0000 UTC" firstStartedPulling="2025-12-06 08:08:42.90199138 +0000 UTC m=+4536.356852424" lastFinishedPulling="2025-12-06 08:08:45.301658809 +0000 UTC m=+4538.756519853" observedRunningTime="2025-12-06 08:08:45.945976552 +0000 UTC m=+4539.400837596" watchObservedRunningTime="2025-12-06 08:08:45.951358166 +0000 UTC m=+4539.406219220" Dec 06 08:08:51 crc kubenswrapper[4945]: I1206 08:08:51.832462 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:51 crc kubenswrapper[4945]: I1206 08:08:51.832804 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:51 crc kubenswrapper[4945]: I1206 08:08:51.886212 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:52 crc kubenswrapper[4945]: I1206 08:08:52.017334 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:55 crc kubenswrapper[4945]: I1206 08:08:55.504367 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gp5hc"] Dec 06 08:08:55 crc kubenswrapper[4945]: I1206 08:08:55.504959 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gp5hc" podUID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerName="registry-server" containerID="cri-o://914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89" gracePeriod=2 Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.004685 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.012825 4945 generic.go:334] "Generic (PLEG): container finished" podID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerID="914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89" exitCode=0 Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.012894 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gp5hc" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.012898 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gp5hc" event={"ID":"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e","Type":"ContainerDied","Data":"914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89"} Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.012992 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gp5hc" event={"ID":"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e","Type":"ContainerDied","Data":"3bd557615f21225a5c072f6fab41758608a61c0cec2a40697bfe1e8904855f54"} Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.013011 4945 scope.go:117] "RemoveContainer" containerID="914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.044428 4945 scope.go:117] "RemoveContainer" containerID="6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.062606 4945 scope.go:117] "RemoveContainer" containerID="0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.087079 4945 scope.go:117] "RemoveContainer" containerID="914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89" Dec 06 08:08:57 crc kubenswrapper[4945]: E1206 08:08:57.087587 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89\": container with ID starting with 914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89 not found: ID does not exist" containerID="914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.087639 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89"} err="failed to get container status \"914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89\": rpc error: code = NotFound desc = could not find container \"914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89\": container with ID starting with 914544f2166a4128b3507a0e42dbe0771c194f10a638572ee0dad92200bf5b89 not found: ID does not exist" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.087669 4945 scope.go:117] "RemoveContainer" containerID="6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb" Dec 06 08:08:57 crc kubenswrapper[4945]: E1206 08:08:57.088096 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb\": container with ID starting with 6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb not found: ID does not exist" containerID="6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.088142 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb"} err="failed to get container status \"6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb\": rpc error: code = NotFound desc = could not find container \"6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb\": container with ID starting with 6f5fa66128ebbfb41bb719484a92755f956f706b7174e13db7ea005177d81ebb not found: ID does not exist" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.088171 4945 scope.go:117] "RemoveContainer" containerID="0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034" Dec 06 08:08:57 crc kubenswrapper[4945]: E1206 08:08:57.088558 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034\": container with ID starting with 0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034 not found: ID does not exist" containerID="0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.088586 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034"} err="failed to get container status \"0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034\": rpc error: code = NotFound desc = could not find container \"0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034\": container with ID starting with 0e79adc5af0e0f94b5c13ecb3486024bf31b16fea8c22cc51ab2d7fef2fbb034 not found: ID does not exist" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.164940 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-utilities\") pod \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.165128 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-catalog-content\") pod \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.165176 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tzb9\" (UniqueName: \"kubernetes.io/projected/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-kube-api-access-9tzb9\") pod \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\" (UID: \"51ce264d-3257-47dd-a0fa-8b67fd3ecb9e\") " Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.171121 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-utilities" (OuterVolumeSpecName: "utilities") pod "51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" (UID: "51ce264d-3257-47dd-a0fa-8b67fd3ecb9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.177510 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-kube-api-access-9tzb9" (OuterVolumeSpecName: "kube-api-access-9tzb9") pod "51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" (UID: "51ce264d-3257-47dd-a0fa-8b67fd3ecb9e"). InnerVolumeSpecName "kube-api-access-9tzb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.213969 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" (UID: "51ce264d-3257-47dd-a0fa-8b67fd3ecb9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.266633 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.266675 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.266689 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tzb9\" (UniqueName: \"kubernetes.io/projected/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e-kube-api-access-9tzb9\") on node \"crc\" DevicePath \"\"" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.347188 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gp5hc"] Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.354857 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gp5hc"] Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.876087 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl"] Dec 06 08:08:57 crc kubenswrapper[4945]: E1206 08:08:57.879956 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerName="registry-server" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.880001 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerName="registry-server" Dec 06 08:08:57 crc kubenswrapper[4945]: E1206 08:08:57.880043 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerName="extract-utilities" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.880053 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerName="extract-utilities" Dec 06 08:08:57 crc kubenswrapper[4945]: E1206 08:08:57.880080 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerName="extract-content" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.880088 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerName="extract-content" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.880574 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" containerName="registry-server" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.882202 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.889911 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.890144 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.890312 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-mpgsx" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.890748 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.891423 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.910028 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl"] Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.986312 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-config\") pod \"dnsmasq-dns-5f6ffc7dc9-rnwsl\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.986354 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4ngs\" (UniqueName: \"kubernetes.io/projected/dd64c59b-c594-49fd-b641-478deb03a4e1-kube-api-access-p4ngs\") pod \"dnsmasq-dns-5f6ffc7dc9-rnwsl\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:57 crc kubenswrapper[4945]: I1206 08:08:57.986382 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-dns-svc\") pod \"dnsmasq-dns-5f6ffc7dc9-rnwsl\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.087332 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-config\") pod \"dnsmasq-dns-5f6ffc7dc9-rnwsl\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.087380 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4ngs\" (UniqueName: \"kubernetes.io/projected/dd64c59b-c594-49fd-b641-478deb03a4e1-kube-api-access-p4ngs\") pod \"dnsmasq-dns-5f6ffc7dc9-rnwsl\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.087400 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-dns-svc\") pod \"dnsmasq-dns-5f6ffc7dc9-rnwsl\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.089014 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-dns-svc\") pod \"dnsmasq-dns-5f6ffc7dc9-rnwsl\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.089374 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-config\") pod \"dnsmasq-dns-5f6ffc7dc9-rnwsl\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.110526 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4ngs\" (UniqueName: \"kubernetes.io/projected/dd64c59b-c594-49fd-b641-478deb03a4e1-kube-api-access-p4ngs\") pod \"dnsmasq-dns-5f6ffc7dc9-rnwsl\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.160687 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68ddc8d76c-kgdk4"] Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.162122 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.172411 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68ddc8d76c-kgdk4"] Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.193744 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4jl9\" (UniqueName: \"kubernetes.io/projected/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-kube-api-access-r4jl9\") pod \"dnsmasq-dns-68ddc8d76c-kgdk4\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.193797 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-dns-svc\") pod \"dnsmasq-dns-68ddc8d76c-kgdk4\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.193826 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-config\") pod \"dnsmasq-dns-68ddc8d76c-kgdk4\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.206756 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.295807 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4jl9\" (UniqueName: \"kubernetes.io/projected/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-kube-api-access-r4jl9\") pod \"dnsmasq-dns-68ddc8d76c-kgdk4\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.295881 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-dns-svc\") pod \"dnsmasq-dns-68ddc8d76c-kgdk4\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.295921 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-config\") pod \"dnsmasq-dns-68ddc8d76c-kgdk4\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.297075 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-config\") pod \"dnsmasq-dns-68ddc8d76c-kgdk4\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.297521 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-dns-svc\") pod \"dnsmasq-dns-68ddc8d76c-kgdk4\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.338783 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4jl9\" (UniqueName: \"kubernetes.io/projected/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-kube-api-access-r4jl9\") pod \"dnsmasq-dns-68ddc8d76c-kgdk4\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.482905 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.683840 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl"] Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.703244 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68ddc8d76c-kgdk4"] Dec 06 08:08:58 crc kubenswrapper[4945]: I1206 08:08:58.972983 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51ce264d-3257-47dd-a0fa-8b67fd3ecb9e" path="/var/lib/kubelet/pods/51ce264d-3257-47dd-a0fa-8b67fd3ecb9e/volumes" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.049685 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.051099 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.053260 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.056170 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.056988 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.057196 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-bnmgz" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.060116 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" event={"ID":"dd64c59b-c594-49fd-b641-478deb03a4e1","Type":"ContainerStarted","Data":"2fdc033c7e183488ce4cf7d3099751efb6f42a6864c274b28dced26883e953f3"} Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.061591 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.065091 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" event={"ID":"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6","Type":"ContainerStarted","Data":"8c34a6e62c57ad7cb299570a6dad476475e1078fa5096b2623380a82ee1ca1cd"} Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.070084 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.214336 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-49d68440-bee0-48c5-a325-d0bac110d066\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.214433 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74b36cc5-dc80-4a7c-a540-667da7fcaba6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.214456 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.214570 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.215486 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psx9h\" (UniqueName: \"kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-kube-api-access-psx9h\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.215560 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74b36cc5-dc80-4a7c-a540-667da7fcaba6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.215592 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.215659 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.215714 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.292131 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.298462 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.301898 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.302192 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.302394 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.302698 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.303114 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6cgd7" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.305545 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.318024 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.318081 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psx9h\" (UniqueName: \"kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-kube-api-access-psx9h\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.318105 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74b36cc5-dc80-4a7c-a540-667da7fcaba6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.318123 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.318162 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.318189 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.318225 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-49d68440-bee0-48c5-a325-d0bac110d066\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.318266 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74b36cc5-dc80-4a7c-a540-667da7fcaba6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.318304 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.318827 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.319532 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.320152 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.320768 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.324755 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74b36cc5-dc80-4a7c-a540-667da7fcaba6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.326256 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.326297 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-49d68440-bee0-48c5-a325-d0bac110d066\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fe6461d66613f9b353416e64c9be64203bdad8b380547e61feaa851f2266e9da/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.326361 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.327123 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74b36cc5-dc80-4a7c-a540-667da7fcaba6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.335670 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psx9h\" (UniqueName: \"kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-kube-api-access-psx9h\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.368371 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-49d68440-bee0-48c5-a325-d0bac110d066\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") pod \"rabbitmq-server-0\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.393838 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.421000 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h4xt\" (UniqueName: \"kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-kube-api-access-5h4xt\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.421146 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.421783 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.421846 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.422081 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.422137 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/11afabe2-0e72-44a5-862b-1a1746bd703c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.422221 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/11afabe2-0e72-44a5-862b-1a1746bd703c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.422968 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.423098 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.524027 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h4xt\" (UniqueName: \"kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-kube-api-access-5h4xt\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.524390 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.524417 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.524439 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.524487 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.524505 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/11afabe2-0e72-44a5-862b-1a1746bd703c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.524525 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/11afabe2-0e72-44a5-862b-1a1746bd703c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.524547 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.524584 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.525842 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.526018 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.526093 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.526949 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.529969 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/11afabe2-0e72-44a5-862b-1a1746bd703c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.530191 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.530214 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5ca2491cce7e023f03f43f9778e110f4075c77efa7fc91b1b6efa413d7a0eaf3/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.532656 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/11afabe2-0e72-44a5-862b-1a1746bd703c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.533149 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.549076 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h4xt\" (UniqueName: \"kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-kube-api-access-5h4xt\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.562631 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") pod \"rabbitmq-cell1-server-0\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.689216 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:08:59 crc kubenswrapper[4945]: I1206 08:08:59.915967 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.025683 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.057973 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.059847 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.063744 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.064018 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-7jrlq" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.064496 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.068886 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.075595 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.101749 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74b36cc5-dc80-4a7c-a540-667da7fcaba6","Type":"ContainerStarted","Data":"ed3f679c74296ddc5c5f76f9c29d91bb62697f07c3481d27e78faa7799233503"} Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.102377 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.138650 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adf8ff2c-2982-4a81-899e-12d36a3c0fef-operator-scripts\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.138703 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/adf8ff2c-2982-4a81-899e-12d36a3c0fef-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.138735 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/adf8ff2c-2982-4a81-899e-12d36a3c0fef-kolla-config\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.138759 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/adf8ff2c-2982-4a81-899e-12d36a3c0fef-config-data-default\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.138800 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2ebf12aa-8762-4bb4-9e02-01a8985c7f19\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ebf12aa-8762-4bb4-9e02-01a8985c7f19\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.138826 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf8ff2c-2982-4a81-899e-12d36a3c0fef-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.138844 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/adf8ff2c-2982-4a81-899e-12d36a3c0fef-config-data-generated\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.138871 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58q5f\" (UniqueName: \"kubernetes.io/projected/adf8ff2c-2982-4a81-899e-12d36a3c0fef-kube-api-access-58q5f\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.191668 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 08:09:00 crc kubenswrapper[4945]: W1206 08:09:00.197707 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11afabe2_0e72_44a5_862b_1a1746bd703c.slice/crio-15837158fc53cd9461ac6faf4bf26ceed12b815af8082c81aa818a7a54ed25db WatchSource:0}: Error finding container 15837158fc53cd9461ac6faf4bf26ceed12b815af8082c81aa818a7a54ed25db: Status 404 returned error can't find the container with id 15837158fc53cd9461ac6faf4bf26ceed12b815af8082c81aa818a7a54ed25db Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.239857 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/adf8ff2c-2982-4a81-899e-12d36a3c0fef-kolla-config\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.239917 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/adf8ff2c-2982-4a81-899e-12d36a3c0fef-config-data-default\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.239961 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2ebf12aa-8762-4bb4-9e02-01a8985c7f19\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ebf12aa-8762-4bb4-9e02-01a8985c7f19\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.239990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf8ff2c-2982-4a81-899e-12d36a3c0fef-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.240005 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/adf8ff2c-2982-4a81-899e-12d36a3c0fef-config-data-generated\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.240037 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58q5f\" (UniqueName: \"kubernetes.io/projected/adf8ff2c-2982-4a81-899e-12d36a3c0fef-kube-api-access-58q5f\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.240070 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adf8ff2c-2982-4a81-899e-12d36a3c0fef-operator-scripts\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.240091 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/adf8ff2c-2982-4a81-899e-12d36a3c0fef-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.240827 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/adf8ff2c-2982-4a81-899e-12d36a3c0fef-kolla-config\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.241578 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/adf8ff2c-2982-4a81-899e-12d36a3c0fef-config-data-generated\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.242401 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/adf8ff2c-2982-4a81-899e-12d36a3c0fef-config-data-default\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.244906 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adf8ff2c-2982-4a81-899e-12d36a3c0fef-operator-scripts\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.253260 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.253339 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2ebf12aa-8762-4bb4-9e02-01a8985c7f19\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ebf12aa-8762-4bb4-9e02-01a8985c7f19\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e9509dfd3dcc2eed1c35af74011d3e0067ed9e65994667262253ad0bfae8a93e/globalmount\"" pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.265511 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf8ff2c-2982-4a81-899e-12d36a3c0fef-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.267945 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/adf8ff2c-2982-4a81-899e-12d36a3c0fef-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.295124 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2ebf12aa-8762-4bb4-9e02-01a8985c7f19\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ebf12aa-8762-4bb4-9e02-01a8985c7f19\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.296853 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58q5f\" (UniqueName: \"kubernetes.io/projected/adf8ff2c-2982-4a81-899e-12d36a3c0fef-kube-api-access-58q5f\") pod \"openstack-galera-0\" (UID: \"adf8ff2c-2982-4a81-899e-12d36a3c0fef\") " pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.416534 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.528538 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.529560 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.531498 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-v2v22" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.532587 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.538423 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.653387 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c37a3a7-2c98-45ce-b22a-2ea765bf240e-config-data\") pod \"memcached-0\" (UID: \"2c37a3a7-2c98-45ce-b22a-2ea765bf240e\") " pod="openstack/memcached-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.653718 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjlxs\" (UniqueName: \"kubernetes.io/projected/2c37a3a7-2c98-45ce-b22a-2ea765bf240e-kube-api-access-vjlxs\") pod \"memcached-0\" (UID: \"2c37a3a7-2c98-45ce-b22a-2ea765bf240e\") " pod="openstack/memcached-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.653739 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2c37a3a7-2c98-45ce-b22a-2ea765bf240e-kolla-config\") pod \"memcached-0\" (UID: \"2c37a3a7-2c98-45ce-b22a-2ea765bf240e\") " pod="openstack/memcached-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.755606 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjlxs\" (UniqueName: \"kubernetes.io/projected/2c37a3a7-2c98-45ce-b22a-2ea765bf240e-kube-api-access-vjlxs\") pod \"memcached-0\" (UID: \"2c37a3a7-2c98-45ce-b22a-2ea765bf240e\") " pod="openstack/memcached-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.755664 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2c37a3a7-2c98-45ce-b22a-2ea765bf240e-kolla-config\") pod \"memcached-0\" (UID: \"2c37a3a7-2c98-45ce-b22a-2ea765bf240e\") " pod="openstack/memcached-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.756593 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c37a3a7-2c98-45ce-b22a-2ea765bf240e-config-data\") pod \"memcached-0\" (UID: \"2c37a3a7-2c98-45ce-b22a-2ea765bf240e\") " pod="openstack/memcached-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.757558 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c37a3a7-2c98-45ce-b22a-2ea765bf240e-config-data\") pod \"memcached-0\" (UID: \"2c37a3a7-2c98-45ce-b22a-2ea765bf240e\") " pod="openstack/memcached-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.758457 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2c37a3a7-2c98-45ce-b22a-2ea765bf240e-kolla-config\") pod \"memcached-0\" (UID: \"2c37a3a7-2c98-45ce-b22a-2ea765bf240e\") " pod="openstack/memcached-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.791787 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjlxs\" (UniqueName: \"kubernetes.io/projected/2c37a3a7-2c98-45ce-b22a-2ea765bf240e-kube-api-access-vjlxs\") pod \"memcached-0\" (UID: \"2c37a3a7-2c98-45ce-b22a-2ea765bf240e\") " pod="openstack/memcached-0" Dec 06 08:09:00 crc kubenswrapper[4945]: I1206 08:09:00.864989 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.023425 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 08:09:01 crc kubenswrapper[4945]: W1206 08:09:01.049033 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadf8ff2c_2982_4a81_899e_12d36a3c0fef.slice/crio-5b8089f3a542c9aadfbde60b793642a088e39090e997c185ebb0e01999a7adf7 WatchSource:0}: Error finding container 5b8089f3a542c9aadfbde60b793642a088e39090e997c185ebb0e01999a7adf7: Status 404 returned error can't find the container with id 5b8089f3a542c9aadfbde60b793642a088e39090e997c185ebb0e01999a7adf7 Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.112051 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"adf8ff2c-2982-4a81-899e-12d36a3c0fef","Type":"ContainerStarted","Data":"5b8089f3a542c9aadfbde60b793642a088e39090e997c185ebb0e01999a7adf7"} Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.116158 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"11afabe2-0e72-44a5-862b-1a1746bd703c","Type":"ContainerStarted","Data":"15837158fc53cd9461ac6faf4bf26ceed12b815af8082c81aa818a7a54ed25db"} Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.367368 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.555873 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.558042 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.562587 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.562866 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-sxljs" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.563095 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.565178 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.565251 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.671021 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.671173 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkbb5\" (UniqueName: \"kubernetes.io/projected/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-kube-api-access-wkbb5\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.671209 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-41806c71-0aa5-46c9-ad9f-a2ac822acd40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41806c71-0aa5-46c9-ad9f-a2ac822acd40\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.671244 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.672205 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.672338 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.672409 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.672487 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.774913 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.774989 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.775081 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.775124 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkbb5\" (UniqueName: \"kubernetes.io/projected/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-kube-api-access-wkbb5\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.776236 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-41806c71-0aa5-46c9-ad9f-a2ac822acd40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41806c71-0aa5-46c9-ad9f-a2ac822acd40\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.776307 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.776342 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.776393 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.776835 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.777444 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.777698 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.781251 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.789604 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.789655 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-41806c71-0aa5-46c9-ad9f-a2ac822acd40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41806c71-0aa5-46c9-ad9f-a2ac822acd40\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d8ac5de6e04a52a5d312d1ae34a8e6e2629aa314fa02de7ede3093dadd8ad3e9/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.790100 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.790212 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.798264 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkbb5\" (UniqueName: \"kubernetes.io/projected/0ea9657b-8247-4130-bbd7-f8c4b2f12b50-kube-api-access-wkbb5\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.823009 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-41806c71-0aa5-46c9-ad9f-a2ac822acd40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41806c71-0aa5-46c9-ad9f-a2ac822acd40\") pod \"openstack-cell1-galera-0\" (UID: \"0ea9657b-8247-4130-bbd7-f8c4b2f12b50\") " pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:01 crc kubenswrapper[4945]: I1206 08:09:01.914889 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:02 crc kubenswrapper[4945]: I1206 08:09:02.130862 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2c37a3a7-2c98-45ce-b22a-2ea765bf240e","Type":"ContainerStarted","Data":"38dd5827e4a99ab298c0f3395537476c86b73c13d7e63d9c7ddf50709b37d7bf"} Dec 06 08:09:02 crc kubenswrapper[4945]: I1206 08:09:02.342019 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 08:09:02 crc kubenswrapper[4945]: W1206 08:09:02.360731 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ea9657b_8247_4130_bbd7_f8c4b2f12b50.slice/crio-588db2694213dc3469fe7827e40468da02a5883c9d6b71ab31fc87bd6b866757 WatchSource:0}: Error finding container 588db2694213dc3469fe7827e40468da02a5883c9d6b71ab31fc87bd6b866757: Status 404 returned error can't find the container with id 588db2694213dc3469fe7827e40468da02a5883c9d6b71ab31fc87bd6b866757 Dec 06 08:09:03 crc kubenswrapper[4945]: I1206 08:09:03.142314 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0ea9657b-8247-4130-bbd7-f8c4b2f12b50","Type":"ContainerStarted","Data":"588db2694213dc3469fe7827e40468da02a5883c9d6b71ab31fc87bd6b866757"} Dec 06 08:09:08 crc kubenswrapper[4945]: I1206 08:09:08.796250 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:09:08 crc kubenswrapper[4945]: I1206 08:09:08.796766 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:09:08 crc kubenswrapper[4945]: I1206 08:09:08.796836 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 08:09:08 crc kubenswrapper[4945]: I1206 08:09:08.797558 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 08:09:08 crc kubenswrapper[4945]: I1206 08:09:08.797635 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" gracePeriod=600 Dec 06 08:09:11 crc kubenswrapper[4945]: I1206 08:09:11.216894 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" exitCode=0 Dec 06 08:09:11 crc kubenswrapper[4945]: I1206 08:09:11.216989 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4"} Dec 06 08:09:11 crc kubenswrapper[4945]: I1206 08:09:11.217840 4945 scope.go:117] "RemoveContainer" containerID="6bd4a6f6505690c4b9d7a9a3b7401b0aa12c4aa77c0b2a5ba56c5fa3f19605bb" Dec 06 08:09:11 crc kubenswrapper[4945]: E1206 08:09:11.613201 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:09:12 crc kubenswrapper[4945]: I1206 08:09:12.227102 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:09:12 crc kubenswrapper[4945]: E1206 08:09:12.227390 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:09:22 crc kubenswrapper[4945]: I1206 08:09:22.953688 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:09:22 crc kubenswrapper[4945]: E1206 08:09:22.954406 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:09:24 crc kubenswrapper[4945]: E1206 08:09:24.028069 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb" Dec 06 08:09:24 crc kubenswrapper[4945]: E1206 08:09:24.028380 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb" Dec 06 08:09:24 crc kubenswrapper[4945]: E1206 08:09:24.028528 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8chc6h5bh56fh546hb7hc8h67h5bchffh577h697h5b5h5bdh59bhf6hf4h558hb5h578h595h5cchfbh644h59ch7fh654h547h587h5cbh5d5h8fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p4ngs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5f6ffc7dc9-rnwsl_openstack(dd64c59b-c594-49fd-b641-478deb03a4e1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 08:09:24 crc kubenswrapper[4945]: E1206 08:09:24.029762 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" podUID="dd64c59b-c594-49fd-b641-478deb03a4e1" Dec 06 08:09:24 crc kubenswrapper[4945]: I1206 08:09:24.318220 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0ea9657b-8247-4130-bbd7-f8c4b2f12b50","Type":"ContainerStarted","Data":"ea5df4e6b2ad4d39b55144f0c9c299902728248711ca5403b9da370d5a0ea327"} Dec 06 08:09:24 crc kubenswrapper[4945]: I1206 08:09:24.320592 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" event={"ID":"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6","Type":"ContainerStarted","Data":"fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019"} Dec 06 08:09:24 crc kubenswrapper[4945]: I1206 08:09:24.322355 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2c37a3a7-2c98-45ce-b22a-2ea765bf240e","Type":"ContainerStarted","Data":"90cbb99daf68ff5fc3fd7988a295c46d1173c5ddee8ada31ce7ac833b44d8f28"} Dec 06 08:09:24 crc kubenswrapper[4945]: I1206 08:09:24.322865 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 06 08:09:24 crc kubenswrapper[4945]: I1206 08:09:24.325850 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"adf8ff2c-2982-4a81-899e-12d36a3c0fef","Type":"ContainerStarted","Data":"50f3186d3be280c962176364bfd2ddf9adee672a09203fdc5cee89fabddba0b8"} Dec 06 08:09:24 crc kubenswrapper[4945]: I1206 08:09:24.458680 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.54146909 podStartE2EDuration="24.458660071s" podCreationTimestamp="2025-12-06 08:09:00 +0000 UTC" firstStartedPulling="2025-12-06 08:09:01.376153188 +0000 UTC m=+4554.831014232" lastFinishedPulling="2025-12-06 08:09:23.293344169 +0000 UTC m=+4576.748205213" observedRunningTime="2025-12-06 08:09:24.454188131 +0000 UTC m=+4577.909049175" watchObservedRunningTime="2025-12-06 08:09:24.458660071 +0000 UTC m=+4577.913521115" Dec 06 08:09:25 crc kubenswrapper[4945]: I1206 08:09:25.334022 4945 generic.go:334] "Generic (PLEG): container finished" podID="dd64c59b-c594-49fd-b641-478deb03a4e1" containerID="81a19f22335c0a825847fd10d07a04260b32c65f7d07bf10ab51431a52bb1f65" exitCode=0 Dec 06 08:09:25 crc kubenswrapper[4945]: I1206 08:09:25.334081 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" event={"ID":"dd64c59b-c594-49fd-b641-478deb03a4e1","Type":"ContainerDied","Data":"81a19f22335c0a825847fd10d07a04260b32c65f7d07bf10ab51431a52bb1f65"} Dec 06 08:09:25 crc kubenswrapper[4945]: I1206 08:09:25.335829 4945 generic.go:334] "Generic (PLEG): container finished" podID="96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" containerID="fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019" exitCode=0 Dec 06 08:09:25 crc kubenswrapper[4945]: I1206 08:09:25.335874 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" event={"ID":"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6","Type":"ContainerDied","Data":"fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019"} Dec 06 08:09:25 crc kubenswrapper[4945]: I1206 08:09:25.335930 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" event={"ID":"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6","Type":"ContainerStarted","Data":"3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747"} Dec 06 08:09:25 crc kubenswrapper[4945]: I1206 08:09:25.336088 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:09:25 crc kubenswrapper[4945]: I1206 08:09:25.384019 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" podStartSLOduration=1.966190447 podStartE2EDuration="27.384003547s" podCreationTimestamp="2025-12-06 08:08:58 +0000 UTC" firstStartedPulling="2025-12-06 08:08:58.707821718 +0000 UTC m=+4552.162682762" lastFinishedPulling="2025-12-06 08:09:24.125634818 +0000 UTC m=+4577.580495862" observedRunningTime="2025-12-06 08:09:25.380972586 +0000 UTC m=+4578.835833620" watchObservedRunningTime="2025-12-06 08:09:25.384003547 +0000 UTC m=+4578.838864591" Dec 06 08:09:26 crc kubenswrapper[4945]: I1206 08:09:26.352323 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"11afabe2-0e72-44a5-862b-1a1746bd703c","Type":"ContainerStarted","Data":"8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7"} Dec 06 08:09:26 crc kubenswrapper[4945]: I1206 08:09:26.355738 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" event={"ID":"dd64c59b-c594-49fd-b641-478deb03a4e1","Type":"ContainerStarted","Data":"4eb1469da16d5f69c05cc13014fc627c668a2c203561ce27641a738a4c0b14e7"} Dec 06 08:09:26 crc kubenswrapper[4945]: I1206 08:09:26.356070 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:09:26 crc kubenswrapper[4945]: I1206 08:09:26.358107 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74b36cc5-dc80-4a7c-a540-667da7fcaba6","Type":"ContainerStarted","Data":"09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823"} Dec 06 08:09:26 crc kubenswrapper[4945]: I1206 08:09:26.409978 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" podStartSLOduration=-9223372007.444828 podStartE2EDuration="29.409948383s" podCreationTimestamp="2025-12-06 08:08:57 +0000 UTC" firstStartedPulling="2025-12-06 08:08:58.68697595 +0000 UTC m=+4552.141836994" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:09:26.406246324 +0000 UTC m=+4579.861107378" watchObservedRunningTime="2025-12-06 08:09:26.409948383 +0000 UTC m=+4579.864809457" Dec 06 08:09:28 crc kubenswrapper[4945]: I1206 08:09:28.370268 4945 generic.go:334] "Generic (PLEG): container finished" podID="0ea9657b-8247-4130-bbd7-f8c4b2f12b50" containerID="ea5df4e6b2ad4d39b55144f0c9c299902728248711ca5403b9da370d5a0ea327" exitCode=0 Dec 06 08:09:28 crc kubenswrapper[4945]: I1206 08:09:28.370501 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0ea9657b-8247-4130-bbd7-f8c4b2f12b50","Type":"ContainerDied","Data":"ea5df4e6b2ad4d39b55144f0c9c299902728248711ca5403b9da370d5a0ea327"} Dec 06 08:09:28 crc kubenswrapper[4945]: I1206 08:09:28.372617 4945 generic.go:334] "Generic (PLEG): container finished" podID="adf8ff2c-2982-4a81-899e-12d36a3c0fef" containerID="50f3186d3be280c962176364bfd2ddf9adee672a09203fdc5cee89fabddba0b8" exitCode=0 Dec 06 08:09:28 crc kubenswrapper[4945]: I1206 08:09:28.372651 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"adf8ff2c-2982-4a81-899e-12d36a3c0fef","Type":"ContainerDied","Data":"50f3186d3be280c962176364bfd2ddf9adee672a09203fdc5cee89fabddba0b8"} Dec 06 08:09:29 crc kubenswrapper[4945]: I1206 08:09:29.381487 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"adf8ff2c-2982-4a81-899e-12d36a3c0fef","Type":"ContainerStarted","Data":"4e605ce2c888ac060fd1ea9218d05e9104923b0fcc513f6f836ddf5b9bbf3506"} Dec 06 08:09:29 crc kubenswrapper[4945]: I1206 08:09:29.384057 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0ea9657b-8247-4130-bbd7-f8c4b2f12b50","Type":"ContainerStarted","Data":"71d10f96df1522ce70088baf0709279e137c9ef4089b9bcb2873ac28693bfd63"} Dec 06 08:09:29 crc kubenswrapper[4945]: I1206 08:09:29.408106 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.475115605 podStartE2EDuration="30.40808474s" podCreationTimestamp="2025-12-06 08:08:59 +0000 UTC" firstStartedPulling="2025-12-06 08:09:01.052605019 +0000 UTC m=+4554.507466063" lastFinishedPulling="2025-12-06 08:09:23.985574154 +0000 UTC m=+4577.440435198" observedRunningTime="2025-12-06 08:09:29.401180106 +0000 UTC m=+4582.856041150" watchObservedRunningTime="2025-12-06 08:09:29.40808474 +0000 UTC m=+4582.862945784" Dec 06 08:09:29 crc kubenswrapper[4945]: I1206 08:09:29.424566 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.730877415 podStartE2EDuration="29.42454656s" podCreationTimestamp="2025-12-06 08:09:00 +0000 UTC" firstStartedPulling="2025-12-06 08:09:02.368186208 +0000 UTC m=+4555.823047252" lastFinishedPulling="2025-12-06 08:09:24.061855353 +0000 UTC m=+4577.516716397" observedRunningTime="2025-12-06 08:09:29.423214275 +0000 UTC m=+4582.878075349" watchObservedRunningTime="2025-12-06 08:09:29.42454656 +0000 UTC m=+4582.879407604" Dec 06 08:09:30 crc kubenswrapper[4945]: I1206 08:09:30.417378 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 06 08:09:30 crc kubenswrapper[4945]: I1206 08:09:30.417812 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 06 08:09:30 crc kubenswrapper[4945]: I1206 08:09:30.866314 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 06 08:09:31 crc kubenswrapper[4945]: I1206 08:09:31.915728 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:31 crc kubenswrapper[4945]: I1206 08:09:31.916103 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:33 crc kubenswrapper[4945]: I1206 08:09:33.208150 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:09:33 crc kubenswrapper[4945]: I1206 08:09:33.486450 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:09:33 crc kubenswrapper[4945]: I1206 08:09:33.558904 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl"] Dec 06 08:09:33 crc kubenswrapper[4945]: I1206 08:09:33.559147 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" podUID="dd64c59b-c594-49fd-b641-478deb03a4e1" containerName="dnsmasq-dns" containerID="cri-o://4eb1469da16d5f69c05cc13014fc627c668a2c203561ce27641a738a4c0b14e7" gracePeriod=10 Dec 06 08:09:33 crc kubenswrapper[4945]: I1206 08:09:33.953393 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:09:33 crc kubenswrapper[4945]: E1206 08:09:33.954005 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:09:36 crc kubenswrapper[4945]: I1206 08:09:36.437826 4945 generic.go:334] "Generic (PLEG): container finished" podID="dd64c59b-c594-49fd-b641-478deb03a4e1" containerID="4eb1469da16d5f69c05cc13014fc627c668a2c203561ce27641a738a4c0b14e7" exitCode=0 Dec 06 08:09:36 crc kubenswrapper[4945]: I1206 08:09:36.437877 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" event={"ID":"dd64c59b-c594-49fd-b641-478deb03a4e1","Type":"ContainerDied","Data":"4eb1469da16d5f69c05cc13014fc627c668a2c203561ce27641a738a4c0b14e7"} Dec 06 08:09:38 crc kubenswrapper[4945]: I1206 08:09:38.207226 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" podUID="dd64c59b-c594-49fd-b641-478deb03a4e1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.235:5353: connect: connection refused" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.027950 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.168100 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4ngs\" (UniqueName: \"kubernetes.io/projected/dd64c59b-c594-49fd-b641-478deb03a4e1-kube-api-access-p4ngs\") pod \"dd64c59b-c594-49fd-b641-478deb03a4e1\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.168778 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-dns-svc\") pod \"dd64c59b-c594-49fd-b641-478deb03a4e1\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.168823 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-config\") pod \"dd64c59b-c594-49fd-b641-478deb03a4e1\" (UID: \"dd64c59b-c594-49fd-b641-478deb03a4e1\") " Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.174455 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd64c59b-c594-49fd-b641-478deb03a4e1-kube-api-access-p4ngs" (OuterVolumeSpecName: "kube-api-access-p4ngs") pod "dd64c59b-c594-49fd-b641-478deb03a4e1" (UID: "dd64c59b-c594-49fd-b641-478deb03a4e1"). InnerVolumeSpecName "kube-api-access-p4ngs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.207701 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-config" (OuterVolumeSpecName: "config") pod "dd64c59b-c594-49fd-b641-478deb03a4e1" (UID: "dd64c59b-c594-49fd-b641-478deb03a4e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.208543 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dd64c59b-c594-49fd-b641-478deb03a4e1" (UID: "dd64c59b-c594-49fd-b641-478deb03a4e1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.270238 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4ngs\" (UniqueName: \"kubernetes.io/projected/dd64c59b-c594-49fd-b641-478deb03a4e1-kube-api-access-p4ngs\") on node \"crc\" DevicePath \"\"" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.270293 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.270304 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd64c59b-c594-49fd-b641-478deb03a4e1-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.482695 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" event={"ID":"dd64c59b-c594-49fd-b641-478deb03a4e1","Type":"ContainerDied","Data":"2fdc033c7e183488ce4cf7d3099751efb6f42a6864c274b28dced26883e953f3"} Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.482758 4945 scope.go:117] "RemoveContainer" containerID="4eb1469da16d5f69c05cc13014fc627c668a2c203561ce27641a738a4c0b14e7" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.482889 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.504661 4945 scope.go:117] "RemoveContainer" containerID="81a19f22335c0a825847fd10d07a04260b32c65f7d07bf10ab51431a52bb1f65" Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.511931 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl"] Dec 06 08:09:41 crc kubenswrapper[4945]: I1206 08:09:41.517125 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f6ffc7dc9-rnwsl"] Dec 06 08:09:42 crc kubenswrapper[4945]: I1206 08:09:42.963081 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd64c59b-c594-49fd-b641-478deb03a4e1" path="/var/lib/kubelet/pods/dd64c59b-c594-49fd-b641-478deb03a4e1/volumes" Dec 06 08:09:43 crc kubenswrapper[4945]: I1206 08:09:43.240306 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:43 crc kubenswrapper[4945]: I1206 08:09:43.306710 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 06 08:09:43 crc kubenswrapper[4945]: I1206 08:09:43.535195 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 06 08:09:43 crc kubenswrapper[4945]: I1206 08:09:43.623369 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 06 08:09:47 crc kubenswrapper[4945]: I1206 08:09:47.952921 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:09:47 crc kubenswrapper[4945]: E1206 08:09:47.953516 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:09:58 crc kubenswrapper[4945]: I1206 08:09:58.608973 4945 generic.go:334] "Generic (PLEG): container finished" podID="74b36cc5-dc80-4a7c-a540-667da7fcaba6" containerID="09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823" exitCode=0 Dec 06 08:09:58 crc kubenswrapper[4945]: I1206 08:09:58.609054 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74b36cc5-dc80-4a7c-a540-667da7fcaba6","Type":"ContainerDied","Data":"09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823"} Dec 06 08:09:58 crc kubenswrapper[4945]: I1206 08:09:58.612492 4945 generic.go:334] "Generic (PLEG): container finished" podID="11afabe2-0e72-44a5-862b-1a1746bd703c" containerID="8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7" exitCode=0 Dec 06 08:09:58 crc kubenswrapper[4945]: I1206 08:09:58.612566 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"11afabe2-0e72-44a5-862b-1a1746bd703c","Type":"ContainerDied","Data":"8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7"} Dec 06 08:09:59 crc kubenswrapper[4945]: I1206 08:09:59.620701 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"11afabe2-0e72-44a5-862b-1a1746bd703c","Type":"ContainerStarted","Data":"5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf"} Dec 06 08:09:59 crc kubenswrapper[4945]: I1206 08:09:59.621236 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:09:59 crc kubenswrapper[4945]: I1206 08:09:59.623503 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74b36cc5-dc80-4a7c-a540-667da7fcaba6","Type":"ContainerStarted","Data":"85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54"} Dec 06 08:09:59 crc kubenswrapper[4945]: I1206 08:09:59.623660 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 08:09:59 crc kubenswrapper[4945]: I1206 08:09:59.643830 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.538051387 podStartE2EDuration="1m1.643813811s" podCreationTimestamp="2025-12-06 08:08:58 +0000 UTC" firstStartedPulling="2025-12-06 08:09:00.225820197 +0000 UTC m=+4553.680681241" lastFinishedPulling="2025-12-06 08:09:23.331582621 +0000 UTC m=+4576.786443665" observedRunningTime="2025-12-06 08:09:59.637972075 +0000 UTC m=+4613.092833149" watchObservedRunningTime="2025-12-06 08:09:59.643813811 +0000 UTC m=+4613.098674855" Dec 06 08:09:59 crc kubenswrapper[4945]: I1206 08:09:59.663252 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.626831448 podStartE2EDuration="1m1.66323146s" podCreationTimestamp="2025-12-06 08:08:58 +0000 UTC" firstStartedPulling="2025-12-06 08:09:00.025461911 +0000 UTC m=+4553.480322955" lastFinishedPulling="2025-12-06 08:09:24.061861923 +0000 UTC m=+4577.516722967" observedRunningTime="2025-12-06 08:09:59.659231333 +0000 UTC m=+4613.114092387" watchObservedRunningTime="2025-12-06 08:09:59.66323146 +0000 UTC m=+4613.118092504" Dec 06 08:10:01 crc kubenswrapper[4945]: I1206 08:10:01.953676 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:10:01 crc kubenswrapper[4945]: E1206 08:10:01.955548 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:10:09 crc kubenswrapper[4945]: I1206 08:10:09.397000 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 08:10:09 crc kubenswrapper[4945]: I1206 08:10:09.691477 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:14 crc kubenswrapper[4945]: I1206 08:10:14.953746 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:10:14 crc kubenswrapper[4945]: E1206 08:10:14.954836 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.329391 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f7f6bbcbf-26ngg"] Dec 06 08:10:16 crc kubenswrapper[4945]: E1206 08:10:16.330136 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd64c59b-c594-49fd-b641-478deb03a4e1" containerName="init" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.330155 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd64c59b-c594-49fd-b641-478deb03a4e1" containerName="init" Dec 06 08:10:16 crc kubenswrapper[4945]: E1206 08:10:16.330172 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd64c59b-c594-49fd-b641-478deb03a4e1" containerName="dnsmasq-dns" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.330180 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd64c59b-c594-49fd-b641-478deb03a4e1" containerName="dnsmasq-dns" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.330393 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd64c59b-c594-49fd-b641-478deb03a4e1" containerName="dnsmasq-dns" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.331375 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.349146 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f7f6bbcbf-26ngg"] Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.387663 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-dns-svc\") pod \"dnsmasq-dns-6f7f6bbcbf-26ngg\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.388057 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-config\") pod \"dnsmasq-dns-6f7f6bbcbf-26ngg\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.388176 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xld47\" (UniqueName: \"kubernetes.io/projected/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-kube-api-access-xld47\") pod \"dnsmasq-dns-6f7f6bbcbf-26ngg\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.489706 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-config\") pod \"dnsmasq-dns-6f7f6bbcbf-26ngg\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.489775 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xld47\" (UniqueName: \"kubernetes.io/projected/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-kube-api-access-xld47\") pod \"dnsmasq-dns-6f7f6bbcbf-26ngg\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.489825 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-dns-svc\") pod \"dnsmasq-dns-6f7f6bbcbf-26ngg\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.490779 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-config\") pod \"dnsmasq-dns-6f7f6bbcbf-26ngg\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.491039 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-dns-svc\") pod \"dnsmasq-dns-6f7f6bbcbf-26ngg\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.508534 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xld47\" (UniqueName: \"kubernetes.io/projected/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-kube-api-access-xld47\") pod \"dnsmasq-dns-6f7f6bbcbf-26ngg\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:16 crc kubenswrapper[4945]: I1206 08:10:16.651751 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:17 crc kubenswrapper[4945]: I1206 08:10:17.088148 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f7f6bbcbf-26ngg"] Dec 06 08:10:17 crc kubenswrapper[4945]: W1206 08:10:17.098705 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30b41eee_ba2d_45a7_8c7b_9c8a9fe1fd41.slice/crio-e78a3741bb35400b2e1d700e10733f4efcead36b3bcb7b9a4331a3d906ddfe55 WatchSource:0}: Error finding container e78a3741bb35400b2e1d700e10733f4efcead36b3bcb7b9a4331a3d906ddfe55: Status 404 returned error can't find the container with id e78a3741bb35400b2e1d700e10733f4efcead36b3bcb7b9a4331a3d906ddfe55 Dec 06 08:10:17 crc kubenswrapper[4945]: I1206 08:10:17.174119 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 08:10:17 crc kubenswrapper[4945]: I1206 08:10:17.798963 4945 generic.go:334] "Generic (PLEG): container finished" podID="30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" containerID="3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262" exitCode=0 Dec 06 08:10:17 crc kubenswrapper[4945]: I1206 08:10:17.799016 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" event={"ID":"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41","Type":"ContainerDied","Data":"3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262"} Dec 06 08:10:17 crc kubenswrapper[4945]: I1206 08:10:17.799263 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" event={"ID":"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41","Type":"ContainerStarted","Data":"e78a3741bb35400b2e1d700e10733f4efcead36b3bcb7b9a4331a3d906ddfe55"} Dec 06 08:10:18 crc kubenswrapper[4945]: I1206 08:10:18.066959 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 08:10:18 crc kubenswrapper[4945]: I1206 08:10:18.808448 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" event={"ID":"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41","Type":"ContainerStarted","Data":"7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36"} Dec 06 08:10:18 crc kubenswrapper[4945]: I1206 08:10:18.809592 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:19 crc kubenswrapper[4945]: I1206 08:10:19.145261 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="74b36cc5-dc80-4a7c-a540-667da7fcaba6" containerName="rabbitmq" containerID="cri-o://85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54" gracePeriod=604799 Dec 06 08:10:19 crc kubenswrapper[4945]: I1206 08:10:19.394531 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="74b36cc5-dc80-4a7c-a540-667da7fcaba6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.237:5672: connect: connection refused" Dec 06 08:10:19 crc kubenswrapper[4945]: I1206 08:10:19.808842 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="11afabe2-0e72-44a5-862b-1a1746bd703c" containerName="rabbitmq" containerID="cri-o://5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf" gracePeriod=604799 Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.670582 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.692950 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" podStartSLOduration=9.69293287 podStartE2EDuration="9.69293287s" podCreationTimestamp="2025-12-06 08:10:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:10:18.834866564 +0000 UTC m=+4632.289727618" watchObservedRunningTime="2025-12-06 08:10:25.69293287 +0000 UTC m=+4639.147793914" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.724983 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-plugins\") pod \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.725177 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74b36cc5-dc80-4a7c-a540-667da7fcaba6-erlang-cookie-secret\") pod \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.725406 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") pod \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.725617 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74b36cc5-dc80-4a7c-a540-667da7fcaba6-pod-info\") pod \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.725405 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "74b36cc5-dc80-4a7c-a540-667da7fcaba6" (UID: "74b36cc5-dc80-4a7c-a540-667da7fcaba6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.725707 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-erlang-cookie\") pod \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.725840 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-server-conf\") pod \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.725868 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-plugins-conf\") pod \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.725889 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psx9h\" (UniqueName: \"kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-kube-api-access-psx9h\") pod \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.725907 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-confd\") pod \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\" (UID: \"74b36cc5-dc80-4a7c-a540-667da7fcaba6\") " Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.726294 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "74b36cc5-dc80-4a7c-a540-667da7fcaba6" (UID: "74b36cc5-dc80-4a7c-a540-667da7fcaba6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.726569 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.726594 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.726739 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "74b36cc5-dc80-4a7c-a540-667da7fcaba6" (UID: "74b36cc5-dc80-4a7c-a540-667da7fcaba6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.744960 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b36cc5-dc80-4a7c-a540-667da7fcaba6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "74b36cc5-dc80-4a7c-a540-667da7fcaba6" (UID: "74b36cc5-dc80-4a7c-a540-667da7fcaba6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.748711 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/74b36cc5-dc80-4a7c-a540-667da7fcaba6-pod-info" (OuterVolumeSpecName: "pod-info") pod "74b36cc5-dc80-4a7c-a540-667da7fcaba6" (UID: "74b36cc5-dc80-4a7c-a540-667da7fcaba6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.752452 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-server-conf" (OuterVolumeSpecName: "server-conf") pod "74b36cc5-dc80-4a7c-a540-667da7fcaba6" (UID: "74b36cc5-dc80-4a7c-a540-667da7fcaba6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.752547 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066" (OuterVolumeSpecName: "persistence") pod "74b36cc5-dc80-4a7c-a540-667da7fcaba6" (UID: "74b36cc5-dc80-4a7c-a540-667da7fcaba6"). InnerVolumeSpecName "pvc-49d68440-bee0-48c5-a325-d0bac110d066". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.754763 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-kube-api-access-psx9h" (OuterVolumeSpecName: "kube-api-access-psx9h") pod "74b36cc5-dc80-4a7c-a540-667da7fcaba6" (UID: "74b36cc5-dc80-4a7c-a540-667da7fcaba6"). InnerVolumeSpecName "kube-api-access-psx9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.818844 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "74b36cc5-dc80-4a7c-a540-667da7fcaba6" (UID: "74b36cc5-dc80-4a7c-a540-667da7fcaba6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.828158 4945 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74b36cc5-dc80-4a7c-a540-667da7fcaba6-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.828187 4945 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.828196 4945 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74b36cc5-dc80-4a7c-a540-667da7fcaba6-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.828208 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psx9h\" (UniqueName: \"kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-kube-api-access-psx9h\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.828217 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74b36cc5-dc80-4a7c-a540-667da7fcaba6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.828239 4945 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74b36cc5-dc80-4a7c-a540-667da7fcaba6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.828299 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-49d68440-bee0-48c5-a325-d0bac110d066\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") on node \"crc\" " Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.847386 4945 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.847578 4945 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-49d68440-bee0-48c5-a325-d0bac110d066" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066") on node "crc" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.857760 4945 generic.go:334] "Generic (PLEG): container finished" podID="74b36cc5-dc80-4a7c-a540-667da7fcaba6" containerID="85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54" exitCode=0 Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.857803 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74b36cc5-dc80-4a7c-a540-667da7fcaba6","Type":"ContainerDied","Data":"85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54"} Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.857829 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74b36cc5-dc80-4a7c-a540-667da7fcaba6","Type":"ContainerDied","Data":"ed3f679c74296ddc5c5f76f9c29d91bb62697f07c3481d27e78faa7799233503"} Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.857827 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.857845 4945 scope.go:117] "RemoveContainer" containerID="85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.884557 4945 scope.go:117] "RemoveContainer" containerID="09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.898644 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.906145 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.911324 4945 scope.go:117] "RemoveContainer" containerID="85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54" Dec 06 08:10:25 crc kubenswrapper[4945]: E1206 08:10:25.911856 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54\": container with ID starting with 85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54 not found: ID does not exist" containerID="85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.911913 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54"} err="failed to get container status \"85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54\": rpc error: code = NotFound desc = could not find container \"85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54\": container with ID starting with 85bfd16f84f16a6d08973ccd099e6cc1906562ebc6528a07609a2bdb8ec8fe54 not found: ID does not exist" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.911945 4945 scope.go:117] "RemoveContainer" containerID="09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823" Dec 06 08:10:25 crc kubenswrapper[4945]: E1206 08:10:25.912258 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823\": container with ID starting with 09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823 not found: ID does not exist" containerID="09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.912343 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823"} err="failed to get container status \"09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823\": rpc error: code = NotFound desc = could not find container \"09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823\": container with ID starting with 09429047590cc3adfc5b85b55ad303852f5a33723e60025b21ad60e262bfa823 not found: ID does not exist" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.920065 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 08:10:25 crc kubenswrapper[4945]: E1206 08:10:25.920400 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b36cc5-dc80-4a7c-a540-667da7fcaba6" containerName="rabbitmq" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.920416 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b36cc5-dc80-4a7c-a540-667da7fcaba6" containerName="rabbitmq" Dec 06 08:10:25 crc kubenswrapper[4945]: E1206 08:10:25.920454 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b36cc5-dc80-4a7c-a540-667da7fcaba6" containerName="setup-container" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.920460 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b36cc5-dc80-4a7c-a540-667da7fcaba6" containerName="setup-container" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.920662 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="74b36cc5-dc80-4a7c-a540-667da7fcaba6" containerName="rabbitmq" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.928448 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.930099 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-bnmgz" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.930323 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.930507 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.930674 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.933154 4945 reconciler_common.go:293] "Volume detached for volume \"pvc-49d68440-bee0-48c5-a325-d0bac110d066\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.933164 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.934248 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 08:10:25 crc kubenswrapper[4945]: I1206 08:10:25.953505 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:10:25 crc kubenswrapper[4945]: E1206 08:10:25.953807 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.034438 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.034486 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.034543 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.034569 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.034724 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.034817 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-49d68440-bee0-48c5-a325-d0bac110d066\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.034856 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.034879 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94rfs\" (UniqueName: \"kubernetes.io/projected/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-kube-api-access-94rfs\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.034900 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.136067 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.136123 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.136150 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.136200 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.136234 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.136291 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.136344 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-49d68440-bee0-48c5-a325-d0bac110d066\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.136371 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.136393 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94rfs\" (UniqueName: \"kubernetes.io/projected/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-kube-api-access-94rfs\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.137642 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.138203 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.138547 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.139129 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.141387 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.141416 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-49d68440-bee0-48c5-a325-d0bac110d066\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fe6461d66613f9b353416e64c9be64203bdad8b380547e61feaa851f2266e9da/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.141861 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.148951 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.151934 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.153577 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94rfs\" (UniqueName: \"kubernetes.io/projected/9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc-kube-api-access-94rfs\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.181730 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-49d68440-bee0-48c5-a325-d0bac110d066\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d68440-bee0-48c5-a325-d0bac110d066\") pod \"rabbitmq-server-0\" (UID: \"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc\") " pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.243681 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.257952 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.342744 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-erlang-cookie\") pod \"11afabe2-0e72-44a5-862b-1a1746bd703c\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.342812 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-plugins\") pod \"11afabe2-0e72-44a5-862b-1a1746bd703c\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.342831 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-confd\") pod \"11afabe2-0e72-44a5-862b-1a1746bd703c\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.342850 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/11afabe2-0e72-44a5-862b-1a1746bd703c-erlang-cookie-secret\") pod \"11afabe2-0e72-44a5-862b-1a1746bd703c\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.342879 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-server-conf\") pod \"11afabe2-0e72-44a5-862b-1a1746bd703c\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.343027 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") pod \"11afabe2-0e72-44a5-862b-1a1746bd703c\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.343095 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-plugins-conf\") pod \"11afabe2-0e72-44a5-862b-1a1746bd703c\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.343144 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h4xt\" (UniqueName: \"kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-kube-api-access-5h4xt\") pod \"11afabe2-0e72-44a5-862b-1a1746bd703c\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.343200 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/11afabe2-0e72-44a5-862b-1a1746bd703c-pod-info\") pod \"11afabe2-0e72-44a5-862b-1a1746bd703c\" (UID: \"11afabe2-0e72-44a5-862b-1a1746bd703c\") " Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.343351 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "11afabe2-0e72-44a5-862b-1a1746bd703c" (UID: "11afabe2-0e72-44a5-862b-1a1746bd703c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.343376 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "11afabe2-0e72-44a5-862b-1a1746bd703c" (UID: "11afabe2-0e72-44a5-862b-1a1746bd703c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.343475 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.343488 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.343738 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "11afabe2-0e72-44a5-862b-1a1746bd703c" (UID: "11afabe2-0e72-44a5-862b-1a1746bd703c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.348188 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11afabe2-0e72-44a5-862b-1a1746bd703c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "11afabe2-0e72-44a5-862b-1a1746bd703c" (UID: "11afabe2-0e72-44a5-862b-1a1746bd703c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.349860 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-kube-api-access-5h4xt" (OuterVolumeSpecName: "kube-api-access-5h4xt") pod "11afabe2-0e72-44a5-862b-1a1746bd703c" (UID: "11afabe2-0e72-44a5-862b-1a1746bd703c"). InnerVolumeSpecName "kube-api-access-5h4xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.350974 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/11afabe2-0e72-44a5-862b-1a1746bd703c-pod-info" (OuterVolumeSpecName: "pod-info") pod "11afabe2-0e72-44a5-862b-1a1746bd703c" (UID: "11afabe2-0e72-44a5-862b-1a1746bd703c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.353661 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091" (OuterVolumeSpecName: "persistence") pod "11afabe2-0e72-44a5-862b-1a1746bd703c" (UID: "11afabe2-0e72-44a5-862b-1a1746bd703c"). InnerVolumeSpecName "pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.371835 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-server-conf" (OuterVolumeSpecName: "server-conf") pod "11afabe2-0e72-44a5-862b-1a1746bd703c" (UID: "11afabe2-0e72-44a5-862b-1a1746bd703c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.441559 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "11afabe2-0e72-44a5-862b-1a1746bd703c" (UID: "11afabe2-0e72-44a5-862b-1a1746bd703c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.445919 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") on node \"crc\" " Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.445957 4945 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.445971 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h4xt\" (UniqueName: \"kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-kube-api-access-5h4xt\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.445984 4945 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/11afabe2-0e72-44a5-862b-1a1746bd703c-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.445994 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/11afabe2-0e72-44a5-862b-1a1746bd703c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.446004 4945 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/11afabe2-0e72-44a5-862b-1a1746bd703c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.446017 4945 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/11afabe2-0e72-44a5-862b-1a1746bd703c-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.466144 4945 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.466313 4945 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091") on node "crc" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.547883 4945 reconciler_common.go:293] "Volume detached for volume \"pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.653458 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.700116 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68ddc8d76c-kgdk4"] Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.700419 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" podUID="96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" containerName="dnsmasq-dns" containerID="cri-o://3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747" gracePeriod=10 Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.767779 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 08:10:26 crc kubenswrapper[4945]: W1206 08:10:26.775319 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a2c4c88_0eb3_40d9_923e_8e6f2f05dedc.slice/crio-3a286cee5bf45567649804defd92ee7f06109b2d9b47c8030ba6a30eeb024d2c WatchSource:0}: Error finding container 3a286cee5bf45567649804defd92ee7f06109b2d9b47c8030ba6a30eeb024d2c: Status 404 returned error can't find the container with id 3a286cee5bf45567649804defd92ee7f06109b2d9b47c8030ba6a30eeb024d2c Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.872206 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc","Type":"ContainerStarted","Data":"3a286cee5bf45567649804defd92ee7f06109b2d9b47c8030ba6a30eeb024d2c"} Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.877533 4945 generic.go:334] "Generic (PLEG): container finished" podID="11afabe2-0e72-44a5-862b-1a1746bd703c" containerID="5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf" exitCode=0 Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.877566 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"11afabe2-0e72-44a5-862b-1a1746bd703c","Type":"ContainerDied","Data":"5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf"} Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.877584 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"11afabe2-0e72-44a5-862b-1a1746bd703c","Type":"ContainerDied","Data":"15837158fc53cd9461ac6faf4bf26ceed12b815af8082c81aa818a7a54ed25db"} Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.877603 4945 scope.go:117] "RemoveContainer" containerID="5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.877643 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.899231 4945 scope.go:117] "RemoveContainer" containerID="8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.912184 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.917594 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.944003 4945 scope.go:117] "RemoveContainer" containerID="5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf" Dec 06 08:10:26 crc kubenswrapper[4945]: E1206 08:10:26.944687 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf\": container with ID starting with 5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf not found: ID does not exist" containerID="5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.944731 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf"} err="failed to get container status \"5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf\": rpc error: code = NotFound desc = could not find container \"5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf\": container with ID starting with 5e05f566432055f908fe5e45d4af8ace860f76b78e7949aee82d1e812ee72fbf not found: ID does not exist" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.944753 4945 scope.go:117] "RemoveContainer" containerID="8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7" Dec 06 08:10:26 crc kubenswrapper[4945]: E1206 08:10:26.945079 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7\": container with ID starting with 8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7 not found: ID does not exist" containerID="8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.945129 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7"} err="failed to get container status \"8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7\": rpc error: code = NotFound desc = could not find container \"8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7\": container with ID starting with 8333201daf9d3e1c6df89ea70f637f3f72c4e90d387a4172d84cfa7837309de7 not found: ID does not exist" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.973154 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11afabe2-0e72-44a5-862b-1a1746bd703c" path="/var/lib/kubelet/pods/11afabe2-0e72-44a5-862b-1a1746bd703c/volumes" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.974034 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74b36cc5-dc80-4a7c-a540-667da7fcaba6" path="/var/lib/kubelet/pods/74b36cc5-dc80-4a7c-a540-667da7fcaba6/volumes" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.974668 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 08:10:26 crc kubenswrapper[4945]: E1206 08:10:26.976801 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11afabe2-0e72-44a5-862b-1a1746bd703c" containerName="setup-container" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.976851 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="11afabe2-0e72-44a5-862b-1a1746bd703c" containerName="setup-container" Dec 06 08:10:26 crc kubenswrapper[4945]: E1206 08:10:26.976914 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11afabe2-0e72-44a5-862b-1a1746bd703c" containerName="rabbitmq" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.976921 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="11afabe2-0e72-44a5-862b-1a1746bd703c" containerName="rabbitmq" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.977341 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="11afabe2-0e72-44a5-862b-1a1746bd703c" containerName="rabbitmq" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.979010 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.979119 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.984192 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6cgd7" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.984441 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.984571 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.984597 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 08:10:26 crc kubenswrapper[4945]: I1206 08:10:26.984611 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.062314 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e36f4855-c747-4b0e-8648-3b7619707dcb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.062401 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.062429 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e36f4855-c747-4b0e-8648-3b7619707dcb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.062476 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e36f4855-c747-4b0e-8648-3b7619707dcb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.062508 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e36f4855-c747-4b0e-8648-3b7619707dcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.062530 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e36f4855-c747-4b0e-8648-3b7619707dcb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.062574 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz85h\" (UniqueName: \"kubernetes.io/projected/e36f4855-c747-4b0e-8648-3b7619707dcb-kube-api-access-lz85h\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.062597 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e36f4855-c747-4b0e-8648-3b7619707dcb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.062688 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e36f4855-c747-4b0e-8648-3b7619707dcb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.164406 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e36f4855-c747-4b0e-8648-3b7619707dcb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.164464 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e36f4855-c747-4b0e-8648-3b7619707dcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.164483 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e36f4855-c747-4b0e-8648-3b7619707dcb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.164516 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz85h\" (UniqueName: \"kubernetes.io/projected/e36f4855-c747-4b0e-8648-3b7619707dcb-kube-api-access-lz85h\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.164536 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e36f4855-c747-4b0e-8648-3b7619707dcb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.164575 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e36f4855-c747-4b0e-8648-3b7619707dcb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.164624 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e36f4855-c747-4b0e-8648-3b7619707dcb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.164660 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.164676 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e36f4855-c747-4b0e-8648-3b7619707dcb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.165230 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e36f4855-c747-4b0e-8648-3b7619707dcb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.165517 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e36f4855-c747-4b0e-8648-3b7619707dcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.166182 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e36f4855-c747-4b0e-8648-3b7619707dcb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.167946 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e36f4855-c747-4b0e-8648-3b7619707dcb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.168986 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.169000 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e36f4855-c747-4b0e-8648-3b7619707dcb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.169020 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5ca2491cce7e023f03f43f9778e110f4075c77efa7fc91b1b6efa413d7a0eaf3/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.169452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e36f4855-c747-4b0e-8648-3b7619707dcb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.170539 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e36f4855-c747-4b0e-8648-3b7619707dcb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.184617 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz85h\" (UniqueName: \"kubernetes.io/projected/e36f4855-c747-4b0e-8648-3b7619707dcb-kube-api-access-lz85h\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.196172 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c360bc3-9d86-4ff9-8885-ceb2a44b3091\") pod \"rabbitmq-cell1-server-0\" (UID: \"e36f4855-c747-4b0e-8648-3b7619707dcb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.299819 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.631631 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.669905 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-dns-svc\") pod \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.670251 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4jl9\" (UniqueName: \"kubernetes.io/projected/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-kube-api-access-r4jl9\") pod \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.671679 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-config\") pod \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\" (UID: \"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6\") " Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.739651 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-kube-api-access-r4jl9" (OuterVolumeSpecName: "kube-api-access-r4jl9") pod "96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" (UID: "96d8f41a-e0ba-4d0d-b218-0feb8eed41c6"). InnerVolumeSpecName "kube-api-access-r4jl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.754861 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 08:10:27 crc kubenswrapper[4945]: W1206 08:10:27.755852 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode36f4855_c747_4b0e_8648_3b7619707dcb.slice/crio-b5cf49d6fddbad14e8216e374dcc6ded002b1937ed0c84f138863130c8f9eaa0 WatchSource:0}: Error finding container b5cf49d6fddbad14e8216e374dcc6ded002b1937ed0c84f138863130c8f9eaa0: Status 404 returned error can't find the container with id b5cf49d6fddbad14e8216e374dcc6ded002b1937ed0c84f138863130c8f9eaa0 Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.760327 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" (UID: "96d8f41a-e0ba-4d0d-b218-0feb8eed41c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.760667 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-config" (OuterVolumeSpecName: "config") pod "96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" (UID: "96d8f41a-e0ba-4d0d-b218-0feb8eed41c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.773674 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.773704 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.773714 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4jl9\" (UniqueName: \"kubernetes.io/projected/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6-kube-api-access-r4jl9\") on node \"crc\" DevicePath \"\"" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.886525 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e36f4855-c747-4b0e-8648-3b7619707dcb","Type":"ContainerStarted","Data":"b5cf49d6fddbad14e8216e374dcc6ded002b1937ed0c84f138863130c8f9eaa0"} Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.888258 4945 generic.go:334] "Generic (PLEG): container finished" podID="96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" containerID="3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747" exitCode=0 Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.888309 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" event={"ID":"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6","Type":"ContainerDied","Data":"3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747"} Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.888378 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" event={"ID":"96d8f41a-e0ba-4d0d-b218-0feb8eed41c6","Type":"ContainerDied","Data":"8c34a6e62c57ad7cb299570a6dad476475e1078fa5096b2623380a82ee1ca1cd"} Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.888399 4945 scope.go:117] "RemoveContainer" containerID="3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.889068 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68ddc8d76c-kgdk4" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.893140 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc","Type":"ContainerStarted","Data":"8ff83d58014863b705daf46af3a0f5a6afff4a298def12b0f4a5064abe90ea47"} Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.924571 4945 scope.go:117] "RemoveContainer" containerID="fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.941175 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68ddc8d76c-kgdk4"] Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.948923 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68ddc8d76c-kgdk4"] Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.950090 4945 scope.go:117] "RemoveContainer" containerID="3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747" Dec 06 08:10:27 crc kubenswrapper[4945]: E1206 08:10:27.950563 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747\": container with ID starting with 3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747 not found: ID does not exist" containerID="3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.950601 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747"} err="failed to get container status \"3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747\": rpc error: code = NotFound desc = could not find container \"3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747\": container with ID starting with 3e5b2a553c5cc9ddb5ced920d8063efad2890f23b03489c6e5f41d667a91b747 not found: ID does not exist" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.950627 4945 scope.go:117] "RemoveContainer" containerID="fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019" Dec 06 08:10:27 crc kubenswrapper[4945]: E1206 08:10:27.950893 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019\": container with ID starting with fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019 not found: ID does not exist" containerID="fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019" Dec 06 08:10:27 crc kubenswrapper[4945]: I1206 08:10:27.950919 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019"} err="failed to get container status \"fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019\": rpc error: code = NotFound desc = could not find container \"fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019\": container with ID starting with fc9df1ac90b41d90ba7cbe9aa1ff3f4c4fbe708590fc72b4c7cad68e3126e019 not found: ID does not exist" Dec 06 08:10:28 crc kubenswrapper[4945]: I1206 08:10:28.964112 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" path="/var/lib/kubelet/pods/96d8f41a-e0ba-4d0d-b218-0feb8eed41c6/volumes" Dec 06 08:10:29 crc kubenswrapper[4945]: I1206 08:10:29.909988 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e36f4855-c747-4b0e-8648-3b7619707dcb","Type":"ContainerStarted","Data":"0fa9173e5560d081279cc9c5f472377533233a522e6a2653edca1189df90cf09"} Dec 06 08:10:37 crc kubenswrapper[4945]: I1206 08:10:37.953652 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:10:37 crc kubenswrapper[4945]: E1206 08:10:37.954380 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:10:49 crc kubenswrapper[4945]: I1206 08:10:49.953172 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:10:49 crc kubenswrapper[4945]: E1206 08:10:49.954029 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:11:00 crc kubenswrapper[4945]: I1206 08:11:00.139399 4945 generic.go:334] "Generic (PLEG): container finished" podID="9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc" containerID="8ff83d58014863b705daf46af3a0f5a6afff4a298def12b0f4a5064abe90ea47" exitCode=0 Dec 06 08:11:00 crc kubenswrapper[4945]: I1206 08:11:00.140037 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc","Type":"ContainerDied","Data":"8ff83d58014863b705daf46af3a0f5a6afff4a298def12b0f4a5064abe90ea47"} Dec 06 08:11:01 crc kubenswrapper[4945]: I1206 08:11:01.147438 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc","Type":"ContainerStarted","Data":"7662c303b364061c35a404943f0375db8640585b4c466f0955980aa0e5cd8a9d"} Dec 06 08:11:01 crc kubenswrapper[4945]: I1206 08:11:01.147939 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 08:11:01 crc kubenswrapper[4945]: I1206 08:11:01.190412 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.190395377 podStartE2EDuration="36.190395377s" podCreationTimestamp="2025-12-06 08:10:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:11:01.186112522 +0000 UTC m=+4674.640973576" watchObservedRunningTime="2025-12-06 08:11:01.190395377 +0000 UTC m=+4674.645256411" Dec 06 08:11:02 crc kubenswrapper[4945]: E1206 08:11:02.040866 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode36f4855_c747_4b0e_8648_3b7619707dcb.slice/crio-0fa9173e5560d081279cc9c5f472377533233a522e6a2653edca1189df90cf09.scope\": RecentStats: unable to find data in memory cache]" Dec 06 08:11:02 crc kubenswrapper[4945]: I1206 08:11:02.953144 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:11:02 crc kubenswrapper[4945]: E1206 08:11:02.953508 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:11:03 crc kubenswrapper[4945]: I1206 08:11:03.160296 4945 generic.go:334] "Generic (PLEG): container finished" podID="e36f4855-c747-4b0e-8648-3b7619707dcb" containerID="0fa9173e5560d081279cc9c5f472377533233a522e6a2653edca1189df90cf09" exitCode=0 Dec 06 08:11:03 crc kubenswrapper[4945]: I1206 08:11:03.160340 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e36f4855-c747-4b0e-8648-3b7619707dcb","Type":"ContainerDied","Data":"0fa9173e5560d081279cc9c5f472377533233a522e6a2653edca1189df90cf09"} Dec 06 08:11:04 crc kubenswrapper[4945]: I1206 08:11:04.170460 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e36f4855-c747-4b0e-8648-3b7619707dcb","Type":"ContainerStarted","Data":"975e98dac3617276611d9fac9ba00645d184e767998a5cb6b3c3afcbdf225ce0"} Dec 06 08:11:04 crc kubenswrapper[4945]: I1206 08:11:04.170699 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:11:04 crc kubenswrapper[4945]: I1206 08:11:04.200056 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.200029651 podStartE2EDuration="38.200029651s" podCreationTimestamp="2025-12-06 08:10:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:11:04.196589079 +0000 UTC m=+4677.651450133" watchObservedRunningTime="2025-12-06 08:11:04.200029651 +0000 UTC m=+4677.654890695" Dec 06 08:11:13 crc kubenswrapper[4945]: I1206 08:11:13.953538 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:11:13 crc kubenswrapper[4945]: E1206 08:11:13.954224 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:11:16 crc kubenswrapper[4945]: I1206 08:11:16.260414 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 08:11:17 crc kubenswrapper[4945]: I1206 08:11:17.302218 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 08:11:24 crc kubenswrapper[4945]: I1206 08:11:24.871118 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Dec 06 08:11:24 crc kubenswrapper[4945]: E1206 08:11:24.871896 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" containerName="init" Dec 06 08:11:24 crc kubenswrapper[4945]: I1206 08:11:24.871907 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" containerName="init" Dec 06 08:11:24 crc kubenswrapper[4945]: E1206 08:11:24.871920 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" containerName="dnsmasq-dns" Dec 06 08:11:24 crc kubenswrapper[4945]: I1206 08:11:24.871926 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" containerName="dnsmasq-dns" Dec 06 08:11:24 crc kubenswrapper[4945]: I1206 08:11:24.872062 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="96d8f41a-e0ba-4d0d-b218-0feb8eed41c6" containerName="dnsmasq-dns" Dec 06 08:11:24 crc kubenswrapper[4945]: I1206 08:11:24.872528 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 06 08:11:24 crc kubenswrapper[4945]: I1206 08:11:24.875122 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gxzcx" Dec 06 08:11:24 crc kubenswrapper[4945]: I1206 08:11:24.885329 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 06 08:11:24 crc kubenswrapper[4945]: I1206 08:11:24.953159 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:11:24 crc kubenswrapper[4945]: I1206 08:11:24.953484 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r42g\" (UniqueName: \"kubernetes.io/projected/daf7d317-5475-4c25-ac9b-0f98c8efd535-kube-api-access-5r42g\") pod \"mariadb-client-1-default\" (UID: \"daf7d317-5475-4c25-ac9b-0f98c8efd535\") " pod="openstack/mariadb-client-1-default" Dec 06 08:11:24 crc kubenswrapper[4945]: E1206 08:11:24.953553 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:11:25 crc kubenswrapper[4945]: I1206 08:11:25.056090 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r42g\" (UniqueName: \"kubernetes.io/projected/daf7d317-5475-4c25-ac9b-0f98c8efd535-kube-api-access-5r42g\") pod \"mariadb-client-1-default\" (UID: \"daf7d317-5475-4c25-ac9b-0f98c8efd535\") " pod="openstack/mariadb-client-1-default" Dec 06 08:11:25 crc kubenswrapper[4945]: I1206 08:11:25.076530 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r42g\" (UniqueName: \"kubernetes.io/projected/daf7d317-5475-4c25-ac9b-0f98c8efd535-kube-api-access-5r42g\") pod \"mariadb-client-1-default\" (UID: \"daf7d317-5475-4c25-ac9b-0f98c8efd535\") " pod="openstack/mariadb-client-1-default" Dec 06 08:11:25 crc kubenswrapper[4945]: I1206 08:11:25.188979 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 06 08:11:25 crc kubenswrapper[4945]: I1206 08:11:25.701009 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 06 08:11:26 crc kubenswrapper[4945]: I1206 08:11:26.335020 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"daf7d317-5475-4c25-ac9b-0f98c8efd535","Type":"ContainerStarted","Data":"bebf065308100064c1147d6c6174c0a2912b8c5eaf4f8b496ba61032ccfa00c6"} Dec 06 08:11:27 crc kubenswrapper[4945]: I1206 08:11:27.343394 4945 generic.go:334] "Generic (PLEG): container finished" podID="daf7d317-5475-4c25-ac9b-0f98c8efd535" containerID="4c5c07e597bb3b32880df7fba31c8a6eb5ccf842e12c452883deeefed88b9738" exitCode=0 Dec 06 08:11:27 crc kubenswrapper[4945]: I1206 08:11:27.343446 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"daf7d317-5475-4c25-ac9b-0f98c8efd535","Type":"ContainerDied","Data":"4c5c07e597bb3b32880df7fba31c8a6eb5ccf842e12c452883deeefed88b9738"} Dec 06 08:11:28 crc kubenswrapper[4945]: I1206 08:11:28.898799 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 06 08:11:28 crc kubenswrapper[4945]: I1206 08:11:28.932268 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_daf7d317-5475-4c25-ac9b-0f98c8efd535/mariadb-client-1-default/0.log" Dec 06 08:11:28 crc kubenswrapper[4945]: I1206 08:11:28.970199 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 06 08:11:28 crc kubenswrapper[4945]: I1206 08:11:28.970240 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.013344 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r42g\" (UniqueName: \"kubernetes.io/projected/daf7d317-5475-4c25-ac9b-0f98c8efd535-kube-api-access-5r42g\") pod \"daf7d317-5475-4c25-ac9b-0f98c8efd535\" (UID: \"daf7d317-5475-4c25-ac9b-0f98c8efd535\") " Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.018446 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf7d317-5475-4c25-ac9b-0f98c8efd535-kube-api-access-5r42g" (OuterVolumeSpecName: "kube-api-access-5r42g") pod "daf7d317-5475-4c25-ac9b-0f98c8efd535" (UID: "daf7d317-5475-4c25-ac9b-0f98c8efd535"). InnerVolumeSpecName "kube-api-access-5r42g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.115175 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r42g\" (UniqueName: \"kubernetes.io/projected/daf7d317-5475-4c25-ac9b-0f98c8efd535-kube-api-access-5r42g\") on node \"crc\" DevicePath \"\"" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.358677 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bebf065308100064c1147d6c6174c0a2912b8c5eaf4f8b496ba61032ccfa00c6" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.358727 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.486302 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Dec 06 08:11:29 crc kubenswrapper[4945]: E1206 08:11:29.486631 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf7d317-5475-4c25-ac9b-0f98c8efd535" containerName="mariadb-client-1-default" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.486647 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf7d317-5475-4c25-ac9b-0f98c8efd535" containerName="mariadb-client-1-default" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.486814 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf7d317-5475-4c25-ac9b-0f98c8efd535" containerName="mariadb-client-1-default" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.487321 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.490668 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gxzcx" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.494873 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.622754 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spv74\" (UniqueName: \"kubernetes.io/projected/12413032-4678-4a2f-8600-d58a32a65147-kube-api-access-spv74\") pod \"mariadb-client-2-default\" (UID: \"12413032-4678-4a2f-8600-d58a32a65147\") " pod="openstack/mariadb-client-2-default" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.723927 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spv74\" (UniqueName: \"kubernetes.io/projected/12413032-4678-4a2f-8600-d58a32a65147-kube-api-access-spv74\") pod \"mariadb-client-2-default\" (UID: \"12413032-4678-4a2f-8600-d58a32a65147\") " pod="openstack/mariadb-client-2-default" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.740848 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spv74\" (UniqueName: \"kubernetes.io/projected/12413032-4678-4a2f-8600-d58a32a65147-kube-api-access-spv74\") pod \"mariadb-client-2-default\" (UID: \"12413032-4678-4a2f-8600-d58a32a65147\") " pod="openstack/mariadb-client-2-default" Dec 06 08:11:29 crc kubenswrapper[4945]: I1206 08:11:29.804253 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 06 08:11:30 crc kubenswrapper[4945]: I1206 08:11:30.321016 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 06 08:11:30 crc kubenswrapper[4945]: I1206 08:11:30.366991 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"12413032-4678-4a2f-8600-d58a32a65147","Type":"ContainerStarted","Data":"da0a0ad073f16f9f7536301752e509a0734855cb493f6f276fed6cb208234817"} Dec 06 08:11:30 crc kubenswrapper[4945]: I1206 08:11:30.963632 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daf7d317-5475-4c25-ac9b-0f98c8efd535" path="/var/lib/kubelet/pods/daf7d317-5475-4c25-ac9b-0f98c8efd535/volumes" Dec 06 08:11:31 crc kubenswrapper[4945]: I1206 08:11:31.384380 4945 generic.go:334] "Generic (PLEG): container finished" podID="12413032-4678-4a2f-8600-d58a32a65147" containerID="27c3b4ab759e4d2a778e6852a9a8686a450d62aec42835ba5dfc9c71aae3b816" exitCode=1 Dec 06 08:11:31 crc kubenswrapper[4945]: I1206 08:11:31.384429 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"12413032-4678-4a2f-8600-d58a32a65147","Type":"ContainerDied","Data":"27c3b4ab759e4d2a778e6852a9a8686a450d62aec42835ba5dfc9c71aae3b816"} Dec 06 08:11:32 crc kubenswrapper[4945]: I1206 08:11:32.757321 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 06 08:11:32 crc kubenswrapper[4945]: I1206 08:11:32.772839 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_12413032-4678-4a2f-8600-d58a32a65147/mariadb-client-2-default/0.log" Dec 06 08:11:32 crc kubenswrapper[4945]: I1206 08:11:32.806749 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 06 08:11:32 crc kubenswrapper[4945]: I1206 08:11:32.814225 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 06 08:11:32 crc kubenswrapper[4945]: I1206 08:11:32.871180 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spv74\" (UniqueName: \"kubernetes.io/projected/12413032-4678-4a2f-8600-d58a32a65147-kube-api-access-spv74\") pod \"12413032-4678-4a2f-8600-d58a32a65147\" (UID: \"12413032-4678-4a2f-8600-d58a32a65147\") " Dec 06 08:11:32 crc kubenswrapper[4945]: I1206 08:11:32.876248 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12413032-4678-4a2f-8600-d58a32a65147-kube-api-access-spv74" (OuterVolumeSpecName: "kube-api-access-spv74") pod "12413032-4678-4a2f-8600-d58a32a65147" (UID: "12413032-4678-4a2f-8600-d58a32a65147"). InnerVolumeSpecName "kube-api-access-spv74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:11:32 crc kubenswrapper[4945]: I1206 08:11:32.962559 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12413032-4678-4a2f-8600-d58a32a65147" path="/var/lib/kubelet/pods/12413032-4678-4a2f-8600-d58a32a65147/volumes" Dec 06 08:11:32 crc kubenswrapper[4945]: I1206 08:11:32.973823 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spv74\" (UniqueName: \"kubernetes.io/projected/12413032-4678-4a2f-8600-d58a32a65147-kube-api-access-spv74\") on node \"crc\" DevicePath \"\"" Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.369414 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Dec 06 08:11:33 crc kubenswrapper[4945]: E1206 08:11:33.371000 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12413032-4678-4a2f-8600-d58a32a65147" containerName="mariadb-client-2-default" Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.371024 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="12413032-4678-4a2f-8600-d58a32a65147" containerName="mariadb-client-2-default" Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.371244 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="12413032-4678-4a2f-8600-d58a32a65147" containerName="mariadb-client-2-default" Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.371780 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.382070 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.422610 4945 scope.go:117] "RemoveContainer" containerID="27c3b4ab759e4d2a778e6852a9a8686a450d62aec42835ba5dfc9c71aae3b816" Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.422632 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.483132 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8ljp\" (UniqueName: \"kubernetes.io/projected/4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a-kube-api-access-r8ljp\") pod \"mariadb-client-1\" (UID: \"4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a\") " pod="openstack/mariadb-client-1" Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.584959 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8ljp\" (UniqueName: \"kubernetes.io/projected/4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a-kube-api-access-r8ljp\") pod \"mariadb-client-1\" (UID: \"4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a\") " pod="openstack/mariadb-client-1" Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.603320 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8ljp\" (UniqueName: \"kubernetes.io/projected/4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a-kube-api-access-r8ljp\") pod \"mariadb-client-1\" (UID: \"4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a\") " pod="openstack/mariadb-client-1" Dec 06 08:11:33 crc kubenswrapper[4945]: I1206 08:11:33.736898 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 06 08:11:34 crc kubenswrapper[4945]: I1206 08:11:34.244688 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 06 08:11:34 crc kubenswrapper[4945]: I1206 08:11:34.432299 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a","Type":"ContainerStarted","Data":"09fd5c232ea35cc84382145b10c2761819eedfdd66f8a547cd6790632598b984"} Dec 06 08:11:35 crc kubenswrapper[4945]: I1206 08:11:35.444966 4945 generic.go:334] "Generic (PLEG): container finished" podID="4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a" containerID="f57cd020670e357aadb69d88a517781b06f62b4934fd009ab0048a7d6154b707" exitCode=0 Dec 06 08:11:35 crc kubenswrapper[4945]: I1206 08:11:35.445073 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a","Type":"ContainerDied","Data":"f57cd020670e357aadb69d88a517781b06f62b4934fd009ab0048a7d6154b707"} Dec 06 08:11:36 crc kubenswrapper[4945]: I1206 08:11:36.772784 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 06 08:11:36 crc kubenswrapper[4945]: I1206 08:11:36.790968 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a/mariadb-client-1/0.log" Dec 06 08:11:36 crc kubenswrapper[4945]: I1206 08:11:36.823114 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Dec 06 08:11:36 crc kubenswrapper[4945]: I1206 08:11:36.829354 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Dec 06 08:11:36 crc kubenswrapper[4945]: I1206 08:11:36.839129 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8ljp\" (UniqueName: \"kubernetes.io/projected/4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a-kube-api-access-r8ljp\") pod \"4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a\" (UID: \"4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a\") " Dec 06 08:11:36 crc kubenswrapper[4945]: I1206 08:11:36.843394 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a-kube-api-access-r8ljp" (OuterVolumeSpecName: "kube-api-access-r8ljp") pod "4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a" (UID: "4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a"). InnerVolumeSpecName "kube-api-access-r8ljp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:11:36 crc kubenswrapper[4945]: I1206 08:11:36.940766 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8ljp\" (UniqueName: \"kubernetes.io/projected/4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a-kube-api-access-r8ljp\") on node \"crc\" DevicePath \"\"" Dec 06 08:11:36 crc kubenswrapper[4945]: I1206 08:11:36.957485 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:11:36 crc kubenswrapper[4945]: E1206 08:11:36.957831 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:11:36 crc kubenswrapper[4945]: I1206 08:11:36.963141 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a" path="/var/lib/kubelet/pods/4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a/volumes" Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.311575 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Dec 06 08:11:37 crc kubenswrapper[4945]: E1206 08:11:37.312243 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a" containerName="mariadb-client-1" Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.312264 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a" containerName="mariadb-client-1" Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.312418 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbf4ccd-c7f4-403d-a2ff-b6d0e06cb54a" containerName="mariadb-client-1" Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.314479 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.319702 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.446410 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7llxl\" (UniqueName: \"kubernetes.io/projected/1eec1dd0-ab43-403e-bf89-73fdaf48ee3b-kube-api-access-7llxl\") pod \"mariadb-client-4-default\" (UID: \"1eec1dd0-ab43-403e-bf89-73fdaf48ee3b\") " pod="openstack/mariadb-client-4-default" Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.462625 4945 scope.go:117] "RemoveContainer" containerID="f57cd020670e357aadb69d88a517781b06f62b4934fd009ab0048a7d6154b707" Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.462675 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.548615 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7llxl\" (UniqueName: \"kubernetes.io/projected/1eec1dd0-ab43-403e-bf89-73fdaf48ee3b-kube-api-access-7llxl\") pod \"mariadb-client-4-default\" (UID: \"1eec1dd0-ab43-403e-bf89-73fdaf48ee3b\") " pod="openstack/mariadb-client-4-default" Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.569061 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7llxl\" (UniqueName: \"kubernetes.io/projected/1eec1dd0-ab43-403e-bf89-73fdaf48ee3b-kube-api-access-7llxl\") pod \"mariadb-client-4-default\" (UID: \"1eec1dd0-ab43-403e-bf89-73fdaf48ee3b\") " pod="openstack/mariadb-client-4-default" Dec 06 08:11:37 crc kubenswrapper[4945]: I1206 08:11:37.634647 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 06 08:11:38 crc kubenswrapper[4945]: I1206 08:11:38.118995 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 06 08:11:38 crc kubenswrapper[4945]: I1206 08:11:38.471193 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"1eec1dd0-ab43-403e-bf89-73fdaf48ee3b","Type":"ContainerStarted","Data":"cf4d41f3552b6cbef44bd18840a6f336a53ffc376a9eab51d8954ddd4131d890"} Dec 06 08:11:38 crc kubenswrapper[4945]: I1206 08:11:38.471230 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"1eec1dd0-ab43-403e-bf89-73fdaf48ee3b","Type":"ContainerStarted","Data":"6bb940a2f488bdb3d4bf656050d44fcd7f2c263349f4a8dd866ac7567197466e"} Dec 06 08:11:38 crc kubenswrapper[4945]: I1206 08:11:38.489501 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-4-default" podStartSLOduration=1.489478911 podStartE2EDuration="1.489478911s" podCreationTimestamp="2025-12-06 08:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:11:38.481403255 +0000 UTC m=+4711.936264309" watchObservedRunningTime="2025-12-06 08:11:38.489478911 +0000 UTC m=+4711.944339955" Dec 06 08:11:39 crc kubenswrapper[4945]: I1206 08:11:39.483384 4945 generic.go:334] "Generic (PLEG): container finished" podID="1eec1dd0-ab43-403e-bf89-73fdaf48ee3b" containerID="cf4d41f3552b6cbef44bd18840a6f336a53ffc376a9eab51d8954ddd4131d890" exitCode=0 Dec 06 08:11:39 crc kubenswrapper[4945]: I1206 08:11:39.483470 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"1eec1dd0-ab43-403e-bf89-73fdaf48ee3b","Type":"ContainerDied","Data":"cf4d41f3552b6cbef44bd18840a6f336a53ffc376a9eab51d8954ddd4131d890"} Dec 06 08:11:40 crc kubenswrapper[4945]: I1206 08:11:40.970351 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 06 08:11:41 crc kubenswrapper[4945]: I1206 08:11:41.007857 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 06 08:11:41 crc kubenswrapper[4945]: I1206 08:11:41.014754 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 06 08:11:41 crc kubenswrapper[4945]: I1206 08:11:41.112361 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7llxl\" (UniqueName: \"kubernetes.io/projected/1eec1dd0-ab43-403e-bf89-73fdaf48ee3b-kube-api-access-7llxl\") pod \"1eec1dd0-ab43-403e-bf89-73fdaf48ee3b\" (UID: \"1eec1dd0-ab43-403e-bf89-73fdaf48ee3b\") " Dec 06 08:11:41 crc kubenswrapper[4945]: I1206 08:11:41.118254 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eec1dd0-ab43-403e-bf89-73fdaf48ee3b-kube-api-access-7llxl" (OuterVolumeSpecName: "kube-api-access-7llxl") pod "1eec1dd0-ab43-403e-bf89-73fdaf48ee3b" (UID: "1eec1dd0-ab43-403e-bf89-73fdaf48ee3b"). InnerVolumeSpecName "kube-api-access-7llxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:11:41 crc kubenswrapper[4945]: I1206 08:11:41.214088 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7llxl\" (UniqueName: \"kubernetes.io/projected/1eec1dd0-ab43-403e-bf89-73fdaf48ee3b-kube-api-access-7llxl\") on node \"crc\" DevicePath \"\"" Dec 06 08:11:41 crc kubenswrapper[4945]: I1206 08:11:41.504626 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bb940a2f488bdb3d4bf656050d44fcd7f2c263349f4a8dd866ac7567197466e" Dec 06 08:11:41 crc kubenswrapper[4945]: I1206 08:11:41.504725 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 06 08:11:42 crc kubenswrapper[4945]: I1206 08:11:42.961233 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eec1dd0-ab43-403e-bf89-73fdaf48ee3b" path="/var/lib/kubelet/pods/1eec1dd0-ab43-403e-bf89-73fdaf48ee3b/volumes" Dec 06 08:11:44 crc kubenswrapper[4945]: I1206 08:11:44.540980 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Dec 06 08:11:44 crc kubenswrapper[4945]: E1206 08:11:44.541882 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eec1dd0-ab43-403e-bf89-73fdaf48ee3b" containerName="mariadb-client-4-default" Dec 06 08:11:44 crc kubenswrapper[4945]: I1206 08:11:44.541901 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eec1dd0-ab43-403e-bf89-73fdaf48ee3b" containerName="mariadb-client-4-default" Dec 06 08:11:44 crc kubenswrapper[4945]: I1206 08:11:44.542911 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eec1dd0-ab43-403e-bf89-73fdaf48ee3b" containerName="mariadb-client-4-default" Dec 06 08:11:44 crc kubenswrapper[4945]: I1206 08:11:44.543938 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 06 08:11:44 crc kubenswrapper[4945]: I1206 08:11:44.546399 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gxzcx" Dec 06 08:11:44 crc kubenswrapper[4945]: I1206 08:11:44.550661 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 06 08:11:44 crc kubenswrapper[4945]: I1206 08:11:44.669751 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5fpt\" (UniqueName: \"kubernetes.io/projected/0438d7ce-6986-49c7-9781-48c512612515-kube-api-access-c5fpt\") pod \"mariadb-client-5-default\" (UID: \"0438d7ce-6986-49c7-9781-48c512612515\") " pod="openstack/mariadb-client-5-default" Dec 06 08:11:44 crc kubenswrapper[4945]: I1206 08:11:44.772424 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5fpt\" (UniqueName: \"kubernetes.io/projected/0438d7ce-6986-49c7-9781-48c512612515-kube-api-access-c5fpt\") pod \"mariadb-client-5-default\" (UID: \"0438d7ce-6986-49c7-9781-48c512612515\") " pod="openstack/mariadb-client-5-default" Dec 06 08:11:44 crc kubenswrapper[4945]: I1206 08:11:44.796631 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5fpt\" (UniqueName: \"kubernetes.io/projected/0438d7ce-6986-49c7-9781-48c512612515-kube-api-access-c5fpt\") pod \"mariadb-client-5-default\" (UID: \"0438d7ce-6986-49c7-9781-48c512612515\") " pod="openstack/mariadb-client-5-default" Dec 06 08:11:44 crc kubenswrapper[4945]: I1206 08:11:44.865874 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 06 08:11:45 crc kubenswrapper[4945]: I1206 08:11:45.387581 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 06 08:11:45 crc kubenswrapper[4945]: W1206 08:11:45.393234 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0438d7ce_6986_49c7_9781_48c512612515.slice/crio-c4470a24a23e19b67c29c4583f57c23ff101fe9969118ab3c2f771718ec1c5b2 WatchSource:0}: Error finding container c4470a24a23e19b67c29c4583f57c23ff101fe9969118ab3c2f771718ec1c5b2: Status 404 returned error can't find the container with id c4470a24a23e19b67c29c4583f57c23ff101fe9969118ab3c2f771718ec1c5b2 Dec 06 08:11:45 crc kubenswrapper[4945]: I1206 08:11:45.539161 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"0438d7ce-6986-49c7-9781-48c512612515","Type":"ContainerStarted","Data":"c4470a24a23e19b67c29c4583f57c23ff101fe9969118ab3c2f771718ec1c5b2"} Dec 06 08:11:46 crc kubenswrapper[4945]: I1206 08:11:46.550365 4945 generic.go:334] "Generic (PLEG): container finished" podID="0438d7ce-6986-49c7-9781-48c512612515" containerID="14a5568c03a6f199e2529308325d625bd697ebbc72a1d77ecdcffd846126a0bd" exitCode=0 Dec 06 08:11:46 crc kubenswrapper[4945]: I1206 08:11:46.550416 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"0438d7ce-6986-49c7-9781-48c512612515","Type":"ContainerDied","Data":"14a5568c03a6f199e2529308325d625bd697ebbc72a1d77ecdcffd846126a0bd"} Dec 06 08:11:47 crc kubenswrapper[4945]: I1206 08:11:47.910914 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 06 08:11:47 crc kubenswrapper[4945]: I1206 08:11:47.929426 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_0438d7ce-6986-49c7-9781-48c512612515/mariadb-client-5-default/0.log" Dec 06 08:11:47 crc kubenswrapper[4945]: I1206 08:11:47.953184 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:11:47 crc kubenswrapper[4945]: E1206 08:11:47.953565 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:11:47 crc kubenswrapper[4945]: I1206 08:11:47.957451 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 06 08:11:47 crc kubenswrapper[4945]: I1206 08:11:47.963503 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.020303 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5fpt\" (UniqueName: \"kubernetes.io/projected/0438d7ce-6986-49c7-9781-48c512612515-kube-api-access-c5fpt\") pod \"0438d7ce-6986-49c7-9781-48c512612515\" (UID: \"0438d7ce-6986-49c7-9781-48c512612515\") " Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.028181 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0438d7ce-6986-49c7-9781-48c512612515-kube-api-access-c5fpt" (OuterVolumeSpecName: "kube-api-access-c5fpt") pod "0438d7ce-6986-49c7-9781-48c512612515" (UID: "0438d7ce-6986-49c7-9781-48c512612515"). InnerVolumeSpecName "kube-api-access-c5fpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.108357 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Dec 06 08:11:48 crc kubenswrapper[4945]: E1206 08:11:48.108688 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0438d7ce-6986-49c7-9781-48c512612515" containerName="mariadb-client-5-default" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.108708 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0438d7ce-6986-49c7-9781-48c512612515" containerName="mariadb-client-5-default" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.108873 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0438d7ce-6986-49c7-9781-48c512612515" containerName="mariadb-client-5-default" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.109347 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.122037 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.122965 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5fpt\" (UniqueName: \"kubernetes.io/projected/0438d7ce-6986-49c7-9781-48c512612515-kube-api-access-c5fpt\") on node \"crc\" DevicePath \"\"" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.223764 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb47k\" (UniqueName: \"kubernetes.io/projected/827d44a8-5e58-447a-aa7a-1d9ec29c94c4-kube-api-access-nb47k\") pod \"mariadb-client-6-default\" (UID: \"827d44a8-5e58-447a-aa7a-1d9ec29c94c4\") " pod="openstack/mariadb-client-6-default" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.422392 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb47k\" (UniqueName: \"kubernetes.io/projected/827d44a8-5e58-447a-aa7a-1d9ec29c94c4-kube-api-access-nb47k\") pod \"mariadb-client-6-default\" (UID: \"827d44a8-5e58-447a-aa7a-1d9ec29c94c4\") " pod="openstack/mariadb-client-6-default" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.450049 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb47k\" (UniqueName: \"kubernetes.io/projected/827d44a8-5e58-447a-aa7a-1d9ec29c94c4-kube-api-access-nb47k\") pod \"mariadb-client-6-default\" (UID: \"827d44a8-5e58-447a-aa7a-1d9ec29c94c4\") " pod="openstack/mariadb-client-6-default" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.590654 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4470a24a23e19b67c29c4583f57c23ff101fe9969118ab3c2f771718ec1c5b2" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.590741 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.735796 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 06 08:11:48 crc kubenswrapper[4945]: I1206 08:11:48.963592 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0438d7ce-6986-49c7-9781-48c512612515" path="/var/lib/kubelet/pods/0438d7ce-6986-49c7-9781-48c512612515/volumes" Dec 06 08:11:49 crc kubenswrapper[4945]: I1206 08:11:49.226975 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 06 08:11:49 crc kubenswrapper[4945]: W1206 08:11:49.232680 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod827d44a8_5e58_447a_aa7a_1d9ec29c94c4.slice/crio-dd55a30aa411c18bddbfe9d0c897ccca700c53e3d23dae6548738bc4cf9c3c84 WatchSource:0}: Error finding container dd55a30aa411c18bddbfe9d0c897ccca700c53e3d23dae6548738bc4cf9c3c84: Status 404 returned error can't find the container with id dd55a30aa411c18bddbfe9d0c897ccca700c53e3d23dae6548738bc4cf9c3c84 Dec 06 08:11:49 crc kubenswrapper[4945]: I1206 08:11:49.598949 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"827d44a8-5e58-447a-aa7a-1d9ec29c94c4","Type":"ContainerStarted","Data":"3f4b7edabf35d1e03e6a921b2107b26b4eda2700b08cbea8040a43d2d030d1e0"} Dec 06 08:11:49 crc kubenswrapper[4945]: I1206 08:11:49.599342 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"827d44a8-5e58-447a-aa7a-1d9ec29c94c4","Type":"ContainerStarted","Data":"dd55a30aa411c18bddbfe9d0c897ccca700c53e3d23dae6548738bc4cf9c3c84"} Dec 06 08:11:50 crc kubenswrapper[4945]: I1206 08:11:50.607850 4945 generic.go:334] "Generic (PLEG): container finished" podID="827d44a8-5e58-447a-aa7a-1d9ec29c94c4" containerID="3f4b7edabf35d1e03e6a921b2107b26b4eda2700b08cbea8040a43d2d030d1e0" exitCode=1 Dec 06 08:11:50 crc kubenswrapper[4945]: I1206 08:11:50.607893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"827d44a8-5e58-447a-aa7a-1d9ec29c94c4","Type":"ContainerDied","Data":"3f4b7edabf35d1e03e6a921b2107b26b4eda2700b08cbea8040a43d2d030d1e0"} Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.034594 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.058442 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_827d44a8-5e58-447a-aa7a-1d9ec29c94c4/mariadb-client-6-default/0.log" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.073319 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb47k\" (UniqueName: \"kubernetes.io/projected/827d44a8-5e58-447a-aa7a-1d9ec29c94c4-kube-api-access-nb47k\") pod \"827d44a8-5e58-447a-aa7a-1d9ec29c94c4\" (UID: \"827d44a8-5e58-447a-aa7a-1d9ec29c94c4\") " Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.091768 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/827d44a8-5e58-447a-aa7a-1d9ec29c94c4-kube-api-access-nb47k" (OuterVolumeSpecName: "kube-api-access-nb47k") pod "827d44a8-5e58-447a-aa7a-1d9ec29c94c4" (UID: "827d44a8-5e58-447a-aa7a-1d9ec29c94c4"). InnerVolumeSpecName "kube-api-access-nb47k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.100817 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.106368 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.175017 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb47k\" (UniqueName: \"kubernetes.io/projected/827d44a8-5e58-447a-aa7a-1d9ec29c94c4-kube-api-access-nb47k\") on node \"crc\" DevicePath \"\"" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.283182 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Dec 06 08:11:52 crc kubenswrapper[4945]: E1206 08:11:52.283589 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="827d44a8-5e58-447a-aa7a-1d9ec29c94c4" containerName="mariadb-client-6-default" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.283609 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="827d44a8-5e58-447a-aa7a-1d9ec29c94c4" containerName="mariadb-client-6-default" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.283768 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="827d44a8-5e58-447a-aa7a-1d9ec29c94c4" containerName="mariadb-client-6-default" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.284275 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.293991 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.377069 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smfpj\" (UniqueName: \"kubernetes.io/projected/c06b1922-cd18-48ea-ab8c-0e4dd580022e-kube-api-access-smfpj\") pod \"mariadb-client-7-default\" (UID: \"c06b1922-cd18-48ea-ab8c-0e4dd580022e\") " pod="openstack/mariadb-client-7-default" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.479173 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smfpj\" (UniqueName: \"kubernetes.io/projected/c06b1922-cd18-48ea-ab8c-0e4dd580022e-kube-api-access-smfpj\") pod \"mariadb-client-7-default\" (UID: \"c06b1922-cd18-48ea-ab8c-0e4dd580022e\") " pod="openstack/mariadb-client-7-default" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.497750 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smfpj\" (UniqueName: \"kubernetes.io/projected/c06b1922-cd18-48ea-ab8c-0e4dd580022e-kube-api-access-smfpj\") pod \"mariadb-client-7-default\" (UID: \"c06b1922-cd18-48ea-ab8c-0e4dd580022e\") " pod="openstack/mariadb-client-7-default" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.605519 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.624467 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd55a30aa411c18bddbfe9d0c897ccca700c53e3d23dae6548738bc4cf9c3c84" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.624536 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 06 08:11:52 crc kubenswrapper[4945]: I1206 08:11:52.984516 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="827d44a8-5e58-447a-aa7a-1d9ec29c94c4" path="/var/lib/kubelet/pods/827d44a8-5e58-447a-aa7a-1d9ec29c94c4/volumes" Dec 06 08:11:53 crc kubenswrapper[4945]: I1206 08:11:53.125433 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 06 08:11:53 crc kubenswrapper[4945]: W1206 08:11:53.136620 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc06b1922_cd18_48ea_ab8c_0e4dd580022e.slice/crio-7e543e99401651ac7ebce5d71fd2bb36bbac247f329b3cffd3b69208db244440 WatchSource:0}: Error finding container 7e543e99401651ac7ebce5d71fd2bb36bbac247f329b3cffd3b69208db244440: Status 404 returned error can't find the container with id 7e543e99401651ac7ebce5d71fd2bb36bbac247f329b3cffd3b69208db244440 Dec 06 08:11:53 crc kubenswrapper[4945]: I1206 08:11:53.633347 4945 generic.go:334] "Generic (PLEG): container finished" podID="c06b1922-cd18-48ea-ab8c-0e4dd580022e" containerID="50ca7d9afaaa496cc1db00cc3dcac5da2d8756f006df2b1b548b5fc2ebbc74bc" exitCode=0 Dec 06 08:11:53 crc kubenswrapper[4945]: I1206 08:11:53.633399 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"c06b1922-cd18-48ea-ab8c-0e4dd580022e","Type":"ContainerDied","Data":"50ca7d9afaaa496cc1db00cc3dcac5da2d8756f006df2b1b548b5fc2ebbc74bc"} Dec 06 08:11:53 crc kubenswrapper[4945]: I1206 08:11:53.633451 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"c06b1922-cd18-48ea-ab8c-0e4dd580022e","Type":"ContainerStarted","Data":"7e543e99401651ac7ebce5d71fd2bb36bbac247f329b3cffd3b69208db244440"} Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.048529 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.069501 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_c06b1922-cd18-48ea-ab8c-0e4dd580022e/mariadb-client-7-default/0.log" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.100212 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.105092 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.132741 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smfpj\" (UniqueName: \"kubernetes.io/projected/c06b1922-cd18-48ea-ab8c-0e4dd580022e-kube-api-access-smfpj\") pod \"c06b1922-cd18-48ea-ab8c-0e4dd580022e\" (UID: \"c06b1922-cd18-48ea-ab8c-0e4dd580022e\") " Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.139235 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c06b1922-cd18-48ea-ab8c-0e4dd580022e-kube-api-access-smfpj" (OuterVolumeSpecName: "kube-api-access-smfpj") pod "c06b1922-cd18-48ea-ab8c-0e4dd580022e" (UID: "c06b1922-cd18-48ea-ab8c-0e4dd580022e"). InnerVolumeSpecName "kube-api-access-smfpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.234729 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smfpj\" (UniqueName: \"kubernetes.io/projected/c06b1922-cd18-48ea-ab8c-0e4dd580022e-kube-api-access-smfpj\") on node \"crc\" DevicePath \"\"" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.297244 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Dec 06 08:11:55 crc kubenswrapper[4945]: E1206 08:11:55.298056 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c06b1922-cd18-48ea-ab8c-0e4dd580022e" containerName="mariadb-client-7-default" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.298102 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c06b1922-cd18-48ea-ab8c-0e4dd580022e" containerName="mariadb-client-7-default" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.298433 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c06b1922-cd18-48ea-ab8c-0e4dd580022e" containerName="mariadb-client-7-default" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.299052 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.302423 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.438789 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw8nb\" (UniqueName: \"kubernetes.io/projected/589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee-kube-api-access-tw8nb\") pod \"mariadb-client-2\" (UID: \"589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee\") " pod="openstack/mariadb-client-2" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.540657 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw8nb\" (UniqueName: \"kubernetes.io/projected/589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee-kube-api-access-tw8nb\") pod \"mariadb-client-2\" (UID: \"589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee\") " pod="openstack/mariadb-client-2" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.559675 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw8nb\" (UniqueName: \"kubernetes.io/projected/589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee-kube-api-access-tw8nb\") pod \"mariadb-client-2\" (UID: \"589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee\") " pod="openstack/mariadb-client-2" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.621293 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.649842 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e543e99401651ac7ebce5d71fd2bb36bbac247f329b3cffd3b69208db244440" Dec 06 08:11:55 crc kubenswrapper[4945]: I1206 08:11:55.649891 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 06 08:11:56 crc kubenswrapper[4945]: I1206 08:11:56.151452 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 06 08:11:56 crc kubenswrapper[4945]: W1206 08:11:56.159909 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod589fd2d8_a2b8_4007_b3b9_ef1aeabe1bee.slice/crio-164fc6982c4b66d48c03187e71c861cd71024d321b8d467ed1725ad810eb0d12 WatchSource:0}: Error finding container 164fc6982c4b66d48c03187e71c861cd71024d321b8d467ed1725ad810eb0d12: Status 404 returned error can't find the container with id 164fc6982c4b66d48c03187e71c861cd71024d321b8d467ed1725ad810eb0d12 Dec 06 08:11:56 crc kubenswrapper[4945]: I1206 08:11:56.660576 4945 generic.go:334] "Generic (PLEG): container finished" podID="589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee" containerID="170bcfacc1df27ed3fc03ecce2e09eab392d39b3f533ff6421b6412dc5a0cd43" exitCode=0 Dec 06 08:11:56 crc kubenswrapper[4945]: I1206 08:11:56.660635 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee","Type":"ContainerDied","Data":"170bcfacc1df27ed3fc03ecce2e09eab392d39b3f533ff6421b6412dc5a0cd43"} Dec 06 08:11:56 crc kubenswrapper[4945]: I1206 08:11:56.660754 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee","Type":"ContainerStarted","Data":"164fc6982c4b66d48c03187e71c861cd71024d321b8d467ed1725ad810eb0d12"} Dec 06 08:11:56 crc kubenswrapper[4945]: I1206 08:11:56.963811 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c06b1922-cd18-48ea-ab8c-0e4dd580022e" path="/var/lib/kubelet/pods/c06b1922-cd18-48ea-ab8c-0e4dd580022e/volumes" Dec 06 08:11:58 crc kubenswrapper[4945]: I1206 08:11:58.048644 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 06 08:11:58 crc kubenswrapper[4945]: I1206 08:11:58.066127 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee/mariadb-client-2/0.log" Dec 06 08:11:58 crc kubenswrapper[4945]: I1206 08:11:58.092705 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Dec 06 08:11:58 crc kubenswrapper[4945]: I1206 08:11:58.094724 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw8nb\" (UniqueName: \"kubernetes.io/projected/589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee-kube-api-access-tw8nb\") pod \"589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee\" (UID: \"589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee\") " Dec 06 08:11:58 crc kubenswrapper[4945]: I1206 08:11:58.099526 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Dec 06 08:11:58 crc kubenswrapper[4945]: I1206 08:11:58.100940 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee-kube-api-access-tw8nb" (OuterVolumeSpecName: "kube-api-access-tw8nb") pod "589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee" (UID: "589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee"). InnerVolumeSpecName "kube-api-access-tw8nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:11:58 crc kubenswrapper[4945]: I1206 08:11:58.197038 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw8nb\" (UniqueName: \"kubernetes.io/projected/589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee-kube-api-access-tw8nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:11:58 crc kubenswrapper[4945]: I1206 08:11:58.690518 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="164fc6982c4b66d48c03187e71c861cd71024d321b8d467ed1725ad810eb0d12" Dec 06 08:11:58 crc kubenswrapper[4945]: I1206 08:11:58.690581 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 06 08:11:58 crc kubenswrapper[4945]: I1206 08:11:58.965070 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee" path="/var/lib/kubelet/pods/589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee/volumes" Dec 06 08:11:59 crc kubenswrapper[4945]: I1206 08:11:59.954456 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:11:59 crc kubenswrapper[4945]: E1206 08:11:59.955030 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:12:12 crc kubenswrapper[4945]: I1206 08:12:12.953710 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:12:12 crc kubenswrapper[4945]: E1206 08:12:12.954454 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:12:23 crc kubenswrapper[4945]: I1206 08:12:23.952952 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:12:23 crc kubenswrapper[4945]: E1206 08:12:23.953628 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:12:24 crc kubenswrapper[4945]: I1206 08:12:24.101611 4945 scope.go:117] "RemoveContainer" containerID="6231ab2b9e48d0041945fc8631173f9a79801396f6f21724b5f1980ca7f7dbce" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.424179 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rsw2t"] Dec 06 08:12:25 crc kubenswrapper[4945]: E1206 08:12:25.424935 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee" containerName="mariadb-client-2" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.424952 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee" containerName="mariadb-client-2" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.425187 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="589fd2d8-a2b8-4007-b3b9-ef1aeabe1bee" containerName="mariadb-client-2" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.428517 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.457166 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rsw2t"] Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.566347 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffhqj\" (UniqueName: \"kubernetes.io/projected/770f0e90-32b6-42e0-9927-4091e7ef90e0-kube-api-access-ffhqj\") pod \"community-operators-rsw2t\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.566441 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-utilities\") pod \"community-operators-rsw2t\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.566497 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-catalog-content\") pod \"community-operators-rsw2t\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.667838 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-catalog-content\") pod \"community-operators-rsw2t\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.667905 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffhqj\" (UniqueName: \"kubernetes.io/projected/770f0e90-32b6-42e0-9927-4091e7ef90e0-kube-api-access-ffhqj\") pod \"community-operators-rsw2t\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.667989 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-utilities\") pod \"community-operators-rsw2t\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.668390 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-utilities\") pod \"community-operators-rsw2t\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.668413 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-catalog-content\") pod \"community-operators-rsw2t\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.687017 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffhqj\" (UniqueName: \"kubernetes.io/projected/770f0e90-32b6-42e0-9927-4091e7ef90e0-kube-api-access-ffhqj\") pod \"community-operators-rsw2t\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:25 crc kubenswrapper[4945]: I1206 08:12:25.750771 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:26 crc kubenswrapper[4945]: I1206 08:12:26.314148 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rsw2t"] Dec 06 08:12:26 crc kubenswrapper[4945]: I1206 08:12:26.906033 4945 generic.go:334] "Generic (PLEG): container finished" podID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerID="bda026eda3f484ad7af95b1b76cc6d6f3ec5b34ba13227fb68558037f77ff050" exitCode=0 Dec 06 08:12:26 crc kubenswrapper[4945]: I1206 08:12:26.906140 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsw2t" event={"ID":"770f0e90-32b6-42e0-9927-4091e7ef90e0","Type":"ContainerDied","Data":"bda026eda3f484ad7af95b1b76cc6d6f3ec5b34ba13227fb68558037f77ff050"} Dec 06 08:12:26 crc kubenswrapper[4945]: I1206 08:12:26.906417 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsw2t" event={"ID":"770f0e90-32b6-42e0-9927-4091e7ef90e0","Type":"ContainerStarted","Data":"53712ace745cdc786679039d01ce95af5cb84133b4e3e6392013d8f90f12b494"} Dec 06 08:12:27 crc kubenswrapper[4945]: I1206 08:12:27.914806 4945 generic.go:334] "Generic (PLEG): container finished" podID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerID="1dbf570d55841972058ff5de1f4e432191d7691628c6478a2b1def0bbe07b5f4" exitCode=0 Dec 06 08:12:27 crc kubenswrapper[4945]: I1206 08:12:27.914901 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsw2t" event={"ID":"770f0e90-32b6-42e0-9927-4091e7ef90e0","Type":"ContainerDied","Data":"1dbf570d55841972058ff5de1f4e432191d7691628c6478a2b1def0bbe07b5f4"} Dec 06 08:12:28 crc kubenswrapper[4945]: I1206 08:12:28.925041 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsw2t" event={"ID":"770f0e90-32b6-42e0-9927-4091e7ef90e0","Type":"ContainerStarted","Data":"74c70199974d75ba433ea27ada64c4362a75ba83d59a6575c4e836dafe86719c"} Dec 06 08:12:35 crc kubenswrapper[4945]: I1206 08:12:35.751594 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:35 crc kubenswrapper[4945]: I1206 08:12:35.752631 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:35 crc kubenswrapper[4945]: I1206 08:12:35.835803 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:35 crc kubenswrapper[4945]: I1206 08:12:35.860389 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rsw2t" podStartSLOduration=9.472046568 podStartE2EDuration="10.86037324s" podCreationTimestamp="2025-12-06 08:12:25 +0000 UTC" firstStartedPulling="2025-12-06 08:12:26.907702494 +0000 UTC m=+4760.362563538" lastFinishedPulling="2025-12-06 08:12:28.296029166 +0000 UTC m=+4761.750890210" observedRunningTime="2025-12-06 08:12:28.947967834 +0000 UTC m=+4762.402828898" watchObservedRunningTime="2025-12-06 08:12:35.86037324 +0000 UTC m=+4769.315234284" Dec 06 08:12:36 crc kubenswrapper[4945]: I1206 08:12:36.093260 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:36 crc kubenswrapper[4945]: I1206 08:12:36.137775 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rsw2t"] Dec 06 08:12:37 crc kubenswrapper[4945]: I1206 08:12:37.994666 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rsw2t" podUID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerName="registry-server" containerID="cri-o://74c70199974d75ba433ea27ada64c4362a75ba83d59a6575c4e836dafe86719c" gracePeriod=2 Dec 06 08:12:38 crc kubenswrapper[4945]: I1206 08:12:38.954503 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:12:38 crc kubenswrapper[4945]: E1206 08:12:38.956045 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.023369 4945 generic.go:334] "Generic (PLEG): container finished" podID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerID="74c70199974d75ba433ea27ada64c4362a75ba83d59a6575c4e836dafe86719c" exitCode=0 Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.023450 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsw2t" event={"ID":"770f0e90-32b6-42e0-9927-4091e7ef90e0","Type":"ContainerDied","Data":"74c70199974d75ba433ea27ada64c4362a75ba83d59a6575c4e836dafe86719c"} Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.515437 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.707859 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-utilities\") pod \"770f0e90-32b6-42e0-9927-4091e7ef90e0\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.707911 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-catalog-content\") pod \"770f0e90-32b6-42e0-9927-4091e7ef90e0\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.707983 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffhqj\" (UniqueName: \"kubernetes.io/projected/770f0e90-32b6-42e0-9927-4091e7ef90e0-kube-api-access-ffhqj\") pod \"770f0e90-32b6-42e0-9927-4091e7ef90e0\" (UID: \"770f0e90-32b6-42e0-9927-4091e7ef90e0\") " Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.708956 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-utilities" (OuterVolumeSpecName: "utilities") pod "770f0e90-32b6-42e0-9927-4091e7ef90e0" (UID: "770f0e90-32b6-42e0-9927-4091e7ef90e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.709385 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.720221 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/770f0e90-32b6-42e0-9927-4091e7ef90e0-kube-api-access-ffhqj" (OuterVolumeSpecName: "kube-api-access-ffhqj") pod "770f0e90-32b6-42e0-9927-4091e7ef90e0" (UID: "770f0e90-32b6-42e0-9927-4091e7ef90e0"). InnerVolumeSpecName "kube-api-access-ffhqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.761199 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "770f0e90-32b6-42e0-9927-4091e7ef90e0" (UID: "770f0e90-32b6-42e0-9927-4091e7ef90e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.810888 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770f0e90-32b6-42e0-9927-4091e7ef90e0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:12:39 crc kubenswrapper[4945]: I1206 08:12:39.810917 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffhqj\" (UniqueName: \"kubernetes.io/projected/770f0e90-32b6-42e0-9927-4091e7ef90e0-kube-api-access-ffhqj\") on node \"crc\" DevicePath \"\"" Dec 06 08:12:40 crc kubenswrapper[4945]: I1206 08:12:40.033878 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsw2t" event={"ID":"770f0e90-32b6-42e0-9927-4091e7ef90e0","Type":"ContainerDied","Data":"53712ace745cdc786679039d01ce95af5cb84133b4e3e6392013d8f90f12b494"} Dec 06 08:12:40 crc kubenswrapper[4945]: I1206 08:12:40.034006 4945 scope.go:117] "RemoveContainer" containerID="74c70199974d75ba433ea27ada64c4362a75ba83d59a6575c4e836dafe86719c" Dec 06 08:12:40 crc kubenswrapper[4945]: I1206 08:12:40.034019 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsw2t" Dec 06 08:12:40 crc kubenswrapper[4945]: I1206 08:12:40.064962 4945 scope.go:117] "RemoveContainer" containerID="1dbf570d55841972058ff5de1f4e432191d7691628c6478a2b1def0bbe07b5f4" Dec 06 08:12:40 crc kubenswrapper[4945]: I1206 08:12:40.083266 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rsw2t"] Dec 06 08:12:40 crc kubenswrapper[4945]: I1206 08:12:40.090899 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rsw2t"] Dec 06 08:12:40 crc kubenswrapper[4945]: I1206 08:12:40.093325 4945 scope.go:117] "RemoveContainer" containerID="bda026eda3f484ad7af95b1b76cc6d6f3ec5b34ba13227fb68558037f77ff050" Dec 06 08:12:40 crc kubenswrapper[4945]: I1206 08:12:40.962546 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="770f0e90-32b6-42e0-9927-4091e7ef90e0" path="/var/lib/kubelet/pods/770f0e90-32b6-42e0-9927-4091e7ef90e0/volumes" Dec 06 08:12:50 crc kubenswrapper[4945]: I1206 08:12:50.953998 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:12:50 crc kubenswrapper[4945]: E1206 08:12:50.954856 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:13:04 crc kubenswrapper[4945]: I1206 08:13:04.953466 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:13:04 crc kubenswrapper[4945]: E1206 08:13:04.954686 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:13:18 crc kubenswrapper[4945]: I1206 08:13:18.953205 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:13:18 crc kubenswrapper[4945]: E1206 08:13:18.954024 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:13:30 crc kubenswrapper[4945]: I1206 08:13:30.956202 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:13:30 crc kubenswrapper[4945]: E1206 08:13:30.958776 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:13:42 crc kubenswrapper[4945]: I1206 08:13:42.953073 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:13:42 crc kubenswrapper[4945]: E1206 08:13:42.953785 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:13:54 crc kubenswrapper[4945]: I1206 08:13:54.954309 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:13:54 crc kubenswrapper[4945]: E1206 08:13:54.955341 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:14:08 crc kubenswrapper[4945]: I1206 08:14:08.952906 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:14:08 crc kubenswrapper[4945]: E1206 08:14:08.953762 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:14:20 crc kubenswrapper[4945]: I1206 08:14:20.954021 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:14:21 crc kubenswrapper[4945]: I1206 08:14:21.820927 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"8e73a93782f4167b2d20799f17eea711a9f4910590603bc3ce04323932093a55"} Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.684805 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h6pp8"] Dec 06 08:14:33 crc kubenswrapper[4945]: E1206 08:14:33.686043 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerName="registry-server" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.686082 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerName="registry-server" Dec 06 08:14:33 crc kubenswrapper[4945]: E1206 08:14:33.686100 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerName="extract-utilities" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.686109 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerName="extract-utilities" Dec 06 08:14:33 crc kubenswrapper[4945]: E1206 08:14:33.686138 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerName="extract-content" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.686146 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerName="extract-content" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.686379 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="770f0e90-32b6-42e0-9927-4091e7ef90e0" containerName="registry-server" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.687646 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.698060 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h6pp8"] Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.747547 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-catalog-content\") pod \"redhat-operators-h6pp8\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.747648 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-utilities\") pod \"redhat-operators-h6pp8\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.748170 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql8cd\" (UniqueName: \"kubernetes.io/projected/65cea3c6-baba-402f-8c02-e7ac7bd7d647-kube-api-access-ql8cd\") pod \"redhat-operators-h6pp8\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.850146 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql8cd\" (UniqueName: \"kubernetes.io/projected/65cea3c6-baba-402f-8c02-e7ac7bd7d647-kube-api-access-ql8cd\") pod \"redhat-operators-h6pp8\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.850254 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-catalog-content\") pod \"redhat-operators-h6pp8\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.850343 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-utilities\") pod \"redhat-operators-h6pp8\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.851225 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-catalog-content\") pod \"redhat-operators-h6pp8\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.851246 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-utilities\") pod \"redhat-operators-h6pp8\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:33 crc kubenswrapper[4945]: I1206 08:14:33.870026 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql8cd\" (UniqueName: \"kubernetes.io/projected/65cea3c6-baba-402f-8c02-e7ac7bd7d647-kube-api-access-ql8cd\") pod \"redhat-operators-h6pp8\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:34 crc kubenswrapper[4945]: I1206 08:14:34.008338 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:34 crc kubenswrapper[4945]: I1206 08:14:34.461966 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h6pp8"] Dec 06 08:14:34 crc kubenswrapper[4945]: I1206 08:14:34.913579 4945 generic.go:334] "Generic (PLEG): container finished" podID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerID="e89c6395511d11d0201e50653e8041c8a2c3c279f8a91f62ca979251b644aba7" exitCode=0 Dec 06 08:14:34 crc kubenswrapper[4945]: I1206 08:14:34.913622 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h6pp8" event={"ID":"65cea3c6-baba-402f-8c02-e7ac7bd7d647","Type":"ContainerDied","Data":"e89c6395511d11d0201e50653e8041c8a2c3c279f8a91f62ca979251b644aba7"} Dec 06 08:14:34 crc kubenswrapper[4945]: I1206 08:14:34.913649 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h6pp8" event={"ID":"65cea3c6-baba-402f-8c02-e7ac7bd7d647","Type":"ContainerStarted","Data":"95186aa02dcf769622abd7da630d26a5dbf570b39ad1303378e1d29b4776aa4c"} Dec 06 08:14:34 crc kubenswrapper[4945]: I1206 08:14:34.915307 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 08:14:35 crc kubenswrapper[4945]: I1206 08:14:35.922603 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h6pp8" event={"ID":"65cea3c6-baba-402f-8c02-e7ac7bd7d647","Type":"ContainerStarted","Data":"2900dfea98d9e5a88cef89b097dddde5ee1e72c1b64444790498166b213ceaf5"} Dec 06 08:14:36 crc kubenswrapper[4945]: I1206 08:14:36.933663 4945 generic.go:334] "Generic (PLEG): container finished" podID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerID="2900dfea98d9e5a88cef89b097dddde5ee1e72c1b64444790498166b213ceaf5" exitCode=0 Dec 06 08:14:36 crc kubenswrapper[4945]: I1206 08:14:36.933760 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h6pp8" event={"ID":"65cea3c6-baba-402f-8c02-e7ac7bd7d647","Type":"ContainerDied","Data":"2900dfea98d9e5a88cef89b097dddde5ee1e72c1b64444790498166b213ceaf5"} Dec 06 08:14:37 crc kubenswrapper[4945]: I1206 08:14:37.944318 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h6pp8" event={"ID":"65cea3c6-baba-402f-8c02-e7ac7bd7d647","Type":"ContainerStarted","Data":"c3282924f398ad35cf933c7535a2a42203018163bb83f6806f661113003ee4c6"} Dec 06 08:14:37 crc kubenswrapper[4945]: I1206 08:14:37.967844 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h6pp8" podStartSLOduration=2.576337885 podStartE2EDuration="4.967824495s" podCreationTimestamp="2025-12-06 08:14:33 +0000 UTC" firstStartedPulling="2025-12-06 08:14:34.914962185 +0000 UTC m=+4888.369823229" lastFinishedPulling="2025-12-06 08:14:37.306448795 +0000 UTC m=+4890.761309839" observedRunningTime="2025-12-06 08:14:37.960802248 +0000 UTC m=+4891.415663312" watchObservedRunningTime="2025-12-06 08:14:37.967824495 +0000 UTC m=+4891.422685549" Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.273843 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.275254 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.277625 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gxzcx" Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.295370 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.352184 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfrlz\" (UniqueName: \"kubernetes.io/projected/c5917d38-82ff-4f41-ae15-61aca63e378c-kube-api-access-gfrlz\") pod \"mariadb-copy-data\" (UID: \"c5917d38-82ff-4f41-ae15-61aca63e378c\") " pod="openstack/mariadb-copy-data" Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.352313 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\") pod \"mariadb-copy-data\" (UID: \"c5917d38-82ff-4f41-ae15-61aca63e378c\") " pod="openstack/mariadb-copy-data" Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.453301 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfrlz\" (UniqueName: \"kubernetes.io/projected/c5917d38-82ff-4f41-ae15-61aca63e378c-kube-api-access-gfrlz\") pod \"mariadb-copy-data\" (UID: \"c5917d38-82ff-4f41-ae15-61aca63e378c\") " pod="openstack/mariadb-copy-data" Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.453365 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\") pod \"mariadb-copy-data\" (UID: \"c5917d38-82ff-4f41-ae15-61aca63e378c\") " pod="openstack/mariadb-copy-data" Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.457461 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.457524 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\") pod \"mariadb-copy-data\" (UID: \"c5917d38-82ff-4f41-ae15-61aca63e378c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/32cd963d0abc1f1ad667858a14dbaf0310983b5a170e849b046fe5913ad9e21f/globalmount\"" pod="openstack/mariadb-copy-data" Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.474240 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfrlz\" (UniqueName: \"kubernetes.io/projected/c5917d38-82ff-4f41-ae15-61aca63e378c-kube-api-access-gfrlz\") pod \"mariadb-copy-data\" (UID: \"c5917d38-82ff-4f41-ae15-61aca63e378c\") " pod="openstack/mariadb-copy-data" Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.484055 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\") pod \"mariadb-copy-data\" (UID: \"c5917d38-82ff-4f41-ae15-61aca63e378c\") " pod="openstack/mariadb-copy-data" Dec 06 08:14:40 crc kubenswrapper[4945]: I1206 08:14:40.630390 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 06 08:14:41 crc kubenswrapper[4945]: I1206 08:14:41.146033 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 06 08:14:41 crc kubenswrapper[4945]: W1206 08:14:41.151015 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5917d38_82ff_4f41_ae15_61aca63e378c.slice/crio-edc0bc4fb87fa11dd58d47562df7bd5238bbb4a4c7be2a8c55895ca7f96c2f03 WatchSource:0}: Error finding container edc0bc4fb87fa11dd58d47562df7bd5238bbb4a4c7be2a8c55895ca7f96c2f03: Status 404 returned error can't find the container with id edc0bc4fb87fa11dd58d47562df7bd5238bbb4a4c7be2a8c55895ca7f96c2f03 Dec 06 08:14:41 crc kubenswrapper[4945]: I1206 08:14:41.971839 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"c5917d38-82ff-4f41-ae15-61aca63e378c","Type":"ContainerStarted","Data":"560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b"} Dec 06 08:14:41 crc kubenswrapper[4945]: I1206 08:14:41.972189 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"c5917d38-82ff-4f41-ae15-61aca63e378c","Type":"ContainerStarted","Data":"edc0bc4fb87fa11dd58d47562df7bd5238bbb4a4c7be2a8c55895ca7f96c2f03"} Dec 06 08:14:44 crc kubenswrapper[4945]: I1206 08:14:44.009318 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:44 crc kubenswrapper[4945]: I1206 08:14:44.009389 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:44 crc kubenswrapper[4945]: I1206 08:14:44.072635 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:44 crc kubenswrapper[4945]: I1206 08:14:44.104979 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=5.104947445 podStartE2EDuration="5.104947445s" podCreationTimestamp="2025-12-06 08:14:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:14:43.003517432 +0000 UTC m=+4896.458378476" watchObservedRunningTime="2025-12-06 08:14:44.104947445 +0000 UTC m=+4897.559808509" Dec 06 08:14:44 crc kubenswrapper[4945]: E1206 08:14:44.722184 4945 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.220:38974->38.129.56.220:39753: write tcp 38.129.56.220:38974->38.129.56.220:39753: write: broken pipe Dec 06 08:14:45 crc kubenswrapper[4945]: I1206 08:14:45.046016 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:45 crc kubenswrapper[4945]: I1206 08:14:45.093939 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h6pp8"] Dec 06 08:14:46 crc kubenswrapper[4945]: I1206 08:14:46.229183 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 06 08:14:46 crc kubenswrapper[4945]: I1206 08:14:46.230254 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 06 08:14:46 crc kubenswrapper[4945]: I1206 08:14:46.241268 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z42qm\" (UniqueName: \"kubernetes.io/projected/49c19753-27f4-4fe2-8af1-18342330c1c6-kube-api-access-z42qm\") pod \"mariadb-client\" (UID: \"49c19753-27f4-4fe2-8af1-18342330c1c6\") " pod="openstack/mariadb-client" Dec 06 08:14:46 crc kubenswrapper[4945]: I1206 08:14:46.247595 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 06 08:14:46 crc kubenswrapper[4945]: I1206 08:14:46.343381 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z42qm\" (UniqueName: \"kubernetes.io/projected/49c19753-27f4-4fe2-8af1-18342330c1c6-kube-api-access-z42qm\") pod \"mariadb-client\" (UID: \"49c19753-27f4-4fe2-8af1-18342330c1c6\") " pod="openstack/mariadb-client" Dec 06 08:14:46 crc kubenswrapper[4945]: I1206 08:14:46.368023 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z42qm\" (UniqueName: \"kubernetes.io/projected/49c19753-27f4-4fe2-8af1-18342330c1c6-kube-api-access-z42qm\") pod \"mariadb-client\" (UID: \"49c19753-27f4-4fe2-8af1-18342330c1c6\") " pod="openstack/mariadb-client" Dec 06 08:14:46 crc kubenswrapper[4945]: I1206 08:14:46.569332 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 06 08:14:46 crc kubenswrapper[4945]: I1206 08:14:46.992523 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 06 08:14:47 crc kubenswrapper[4945]: I1206 08:14:47.007901 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"49c19753-27f4-4fe2-8af1-18342330c1c6","Type":"ContainerStarted","Data":"5416aafe4b39cd9fccdaf9a4316e78bdfcfbd78e511bf37eb5a2e1d28d2e0a60"} Dec 06 08:14:47 crc kubenswrapper[4945]: I1206 08:14:47.008144 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h6pp8" podUID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerName="registry-server" containerID="cri-o://c3282924f398ad35cf933c7535a2a42203018163bb83f6806f661113003ee4c6" gracePeriod=2 Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.026733 4945 generic.go:334] "Generic (PLEG): container finished" podID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerID="c3282924f398ad35cf933c7535a2a42203018163bb83f6806f661113003ee4c6" exitCode=0 Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.026845 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h6pp8" event={"ID":"65cea3c6-baba-402f-8c02-e7ac7bd7d647","Type":"ContainerDied","Data":"c3282924f398ad35cf933c7535a2a42203018163bb83f6806f661113003ee4c6"} Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.032345 4945 generic.go:334] "Generic (PLEG): container finished" podID="49c19753-27f4-4fe2-8af1-18342330c1c6" containerID="0ebcd95312b81ff60125becbb521be435fd9f06c6b38d4db854a827e909d016e" exitCode=0 Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.032417 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"49c19753-27f4-4fe2-8af1-18342330c1c6","Type":"ContainerDied","Data":"0ebcd95312b81ff60125becbb521be435fd9f06c6b38d4db854a827e909d016e"} Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.290215 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.391938 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql8cd\" (UniqueName: \"kubernetes.io/projected/65cea3c6-baba-402f-8c02-e7ac7bd7d647-kube-api-access-ql8cd\") pod \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.392037 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-utilities\") pod \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.392108 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-catalog-content\") pod \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\" (UID: \"65cea3c6-baba-402f-8c02-e7ac7bd7d647\") " Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.393064 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-utilities" (OuterVolumeSpecName: "utilities") pod "65cea3c6-baba-402f-8c02-e7ac7bd7d647" (UID: "65cea3c6-baba-402f-8c02-e7ac7bd7d647"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.397780 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65cea3c6-baba-402f-8c02-e7ac7bd7d647-kube-api-access-ql8cd" (OuterVolumeSpecName: "kube-api-access-ql8cd") pod "65cea3c6-baba-402f-8c02-e7ac7bd7d647" (UID: "65cea3c6-baba-402f-8c02-e7ac7bd7d647"). InnerVolumeSpecName "kube-api-access-ql8cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.493491 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql8cd\" (UniqueName: \"kubernetes.io/projected/65cea3c6-baba-402f-8c02-e7ac7bd7d647-kube-api-access-ql8cd\") on node \"crc\" DevicePath \"\"" Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.493531 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.506054 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65cea3c6-baba-402f-8c02-e7ac7bd7d647" (UID: "65cea3c6-baba-402f-8c02-e7ac7bd7d647"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:14:49 crc kubenswrapper[4945]: I1206 08:14:49.596101 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65cea3c6-baba-402f-8c02-e7ac7bd7d647-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.043588 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h6pp8" event={"ID":"65cea3c6-baba-402f-8c02-e7ac7bd7d647","Type":"ContainerDied","Data":"95186aa02dcf769622abd7da630d26a5dbf570b39ad1303378e1d29b4776aa4c"} Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.043675 4945 scope.go:117] "RemoveContainer" containerID="c3282924f398ad35cf933c7535a2a42203018163bb83f6806f661113003ee4c6" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.043601 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h6pp8" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.067739 4945 scope.go:117] "RemoveContainer" containerID="2900dfea98d9e5a88cef89b097dddde5ee1e72c1b64444790498166b213ceaf5" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.086535 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h6pp8"] Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.100382 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h6pp8"] Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.109203 4945 scope.go:117] "RemoveContainer" containerID="e89c6395511d11d0201e50653e8041c8a2c3c279f8a91f62ca979251b644aba7" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.349893 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.369917 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_49c19753-27f4-4fe2-8af1-18342330c1c6/mariadb-client/0.log" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.399801 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.408330 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.509253 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z42qm\" (UniqueName: \"kubernetes.io/projected/49c19753-27f4-4fe2-8af1-18342330c1c6-kube-api-access-z42qm\") pod \"49c19753-27f4-4fe2-8af1-18342330c1c6\" (UID: \"49c19753-27f4-4fe2-8af1-18342330c1c6\") " Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.514182 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c19753-27f4-4fe2-8af1-18342330c1c6-kube-api-access-z42qm" (OuterVolumeSpecName: "kube-api-access-z42qm") pod "49c19753-27f4-4fe2-8af1-18342330c1c6" (UID: "49c19753-27f4-4fe2-8af1-18342330c1c6"). InnerVolumeSpecName "kube-api-access-z42qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.581639 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 06 08:14:50 crc kubenswrapper[4945]: E1206 08:14:50.581976 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49c19753-27f4-4fe2-8af1-18342330c1c6" containerName="mariadb-client" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.581993 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="49c19753-27f4-4fe2-8af1-18342330c1c6" containerName="mariadb-client" Dec 06 08:14:50 crc kubenswrapper[4945]: E1206 08:14:50.582004 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerName="extract-content" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.582010 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerName="extract-content" Dec 06 08:14:50 crc kubenswrapper[4945]: E1206 08:14:50.582041 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerName="registry-server" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.582046 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerName="registry-server" Dec 06 08:14:50 crc kubenswrapper[4945]: E1206 08:14:50.582057 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerName="extract-utilities" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.582065 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerName="extract-utilities" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.582231 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" containerName="registry-server" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.582246 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="49c19753-27f4-4fe2-8af1-18342330c1c6" containerName="mariadb-client" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.582787 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.586412 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.611070 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z42qm\" (UniqueName: \"kubernetes.io/projected/49c19753-27f4-4fe2-8af1-18342330c1c6-kube-api-access-z42qm\") on node \"crc\" DevicePath \"\"" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.712547 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t86sc\" (UniqueName: \"kubernetes.io/projected/8f5c28ae-e992-48f6-9366-6709c12db25e-kube-api-access-t86sc\") pod \"mariadb-client\" (UID: \"8f5c28ae-e992-48f6-9366-6709c12db25e\") " pod="openstack/mariadb-client" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.814355 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t86sc\" (UniqueName: \"kubernetes.io/projected/8f5c28ae-e992-48f6-9366-6709c12db25e-kube-api-access-t86sc\") pod \"mariadb-client\" (UID: \"8f5c28ae-e992-48f6-9366-6709c12db25e\") " pod="openstack/mariadb-client" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.831915 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t86sc\" (UniqueName: \"kubernetes.io/projected/8f5c28ae-e992-48f6-9366-6709c12db25e-kube-api-access-t86sc\") pod \"mariadb-client\" (UID: \"8f5c28ae-e992-48f6-9366-6709c12db25e\") " pod="openstack/mariadb-client" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.909160 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.962138 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c19753-27f4-4fe2-8af1-18342330c1c6" path="/var/lib/kubelet/pods/49c19753-27f4-4fe2-8af1-18342330c1c6/volumes" Dec 06 08:14:50 crc kubenswrapper[4945]: I1206 08:14:50.962968 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65cea3c6-baba-402f-8c02-e7ac7bd7d647" path="/var/lib/kubelet/pods/65cea3c6-baba-402f-8c02-e7ac7bd7d647/volumes" Dec 06 08:14:51 crc kubenswrapper[4945]: I1206 08:14:51.052861 4945 scope.go:117] "RemoveContainer" containerID="0ebcd95312b81ff60125becbb521be435fd9f06c6b38d4db854a827e909d016e" Dec 06 08:14:51 crc kubenswrapper[4945]: I1206 08:14:51.052888 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 06 08:14:51 crc kubenswrapper[4945]: I1206 08:14:51.314165 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 06 08:14:51 crc kubenswrapper[4945]: W1206 08:14:51.318702 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f5c28ae_e992_48f6_9366_6709c12db25e.slice/crio-4baefa856d3f093539354da687bc4cc1c90087ba85e23ded8da6664ee185da57 WatchSource:0}: Error finding container 4baefa856d3f093539354da687bc4cc1c90087ba85e23ded8da6664ee185da57: Status 404 returned error can't find the container with id 4baefa856d3f093539354da687bc4cc1c90087ba85e23ded8da6664ee185da57 Dec 06 08:14:52 crc kubenswrapper[4945]: I1206 08:14:52.062637 4945 generic.go:334] "Generic (PLEG): container finished" podID="8f5c28ae-e992-48f6-9366-6709c12db25e" containerID="428b11892998719a52f7cde2735cda8e80c06df12c4b80e3e4fdd817422035d6" exitCode=0 Dec 06 08:14:52 crc kubenswrapper[4945]: I1206 08:14:52.063110 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"8f5c28ae-e992-48f6-9366-6709c12db25e","Type":"ContainerDied","Data":"428b11892998719a52f7cde2735cda8e80c06df12c4b80e3e4fdd817422035d6"} Dec 06 08:14:52 crc kubenswrapper[4945]: I1206 08:14:52.063140 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"8f5c28ae-e992-48f6-9366-6709c12db25e","Type":"ContainerStarted","Data":"4baefa856d3f093539354da687bc4cc1c90087ba85e23ded8da6664ee185da57"} Dec 06 08:14:53 crc kubenswrapper[4945]: I1206 08:14:53.400929 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 06 08:14:53 crc kubenswrapper[4945]: I1206 08:14:53.420201 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_8f5c28ae-e992-48f6-9366-6709c12db25e/mariadb-client/0.log" Dec 06 08:14:53 crc kubenswrapper[4945]: I1206 08:14:53.447420 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 06 08:14:53 crc kubenswrapper[4945]: I1206 08:14:53.453096 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 06 08:14:53 crc kubenswrapper[4945]: I1206 08:14:53.555066 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t86sc\" (UniqueName: \"kubernetes.io/projected/8f5c28ae-e992-48f6-9366-6709c12db25e-kube-api-access-t86sc\") pod \"8f5c28ae-e992-48f6-9366-6709c12db25e\" (UID: \"8f5c28ae-e992-48f6-9366-6709c12db25e\") " Dec 06 08:14:53 crc kubenswrapper[4945]: I1206 08:14:53.564951 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f5c28ae-e992-48f6-9366-6709c12db25e-kube-api-access-t86sc" (OuterVolumeSpecName: "kube-api-access-t86sc") pod "8f5c28ae-e992-48f6-9366-6709c12db25e" (UID: "8f5c28ae-e992-48f6-9366-6709c12db25e"). InnerVolumeSpecName "kube-api-access-t86sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:14:53 crc kubenswrapper[4945]: I1206 08:14:53.657072 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t86sc\" (UniqueName: \"kubernetes.io/projected/8f5c28ae-e992-48f6-9366-6709c12db25e-kube-api-access-t86sc\") on node \"crc\" DevicePath \"\"" Dec 06 08:14:54 crc kubenswrapper[4945]: I1206 08:14:54.080816 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4baefa856d3f093539354da687bc4cc1c90087ba85e23ded8da6664ee185da57" Dec 06 08:14:54 crc kubenswrapper[4945]: I1206 08:14:54.080880 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 06 08:14:54 crc kubenswrapper[4945]: I1206 08:14:54.966034 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f5c28ae-e992-48f6-9366-6709c12db25e" path="/var/lib/kubelet/pods/8f5c28ae-e992-48f6-9366-6709c12db25e/volumes" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.163948 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j"] Dec 06 08:15:00 crc kubenswrapper[4945]: E1206 08:15:00.165751 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f5c28ae-e992-48f6-9366-6709c12db25e" containerName="mariadb-client" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.165785 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f5c28ae-e992-48f6-9366-6709c12db25e" containerName="mariadb-client" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.166197 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f5c28ae-e992-48f6-9366-6709c12db25e" containerName="mariadb-client" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.167526 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.171058 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.171545 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.175169 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-secret-volume\") pod \"collect-profiles-29416815-8x25j\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.175262 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp8f9\" (UniqueName: \"kubernetes.io/projected/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-kube-api-access-qp8f9\") pod \"collect-profiles-29416815-8x25j\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.175628 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-config-volume\") pod \"collect-profiles-29416815-8x25j\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.176582 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j"] Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.276984 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-secret-volume\") pod \"collect-profiles-29416815-8x25j\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.277061 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp8f9\" (UniqueName: \"kubernetes.io/projected/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-kube-api-access-qp8f9\") pod \"collect-profiles-29416815-8x25j\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.277115 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-config-volume\") pod \"collect-profiles-29416815-8x25j\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.278209 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-config-volume\") pod \"collect-profiles-29416815-8x25j\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.283957 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-secret-volume\") pod \"collect-profiles-29416815-8x25j\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.295208 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp8f9\" (UniqueName: \"kubernetes.io/projected/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-kube-api-access-qp8f9\") pod \"collect-profiles-29416815-8x25j\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.506201 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:00 crc kubenswrapper[4945]: I1206 08:15:00.917022 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j"] Dec 06 08:15:01 crc kubenswrapper[4945]: I1206 08:15:01.154529 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" event={"ID":"e22d9737-bf91-4dd2-8262-6e43b23ff1d2","Type":"ContainerStarted","Data":"d310a26a2941cadf470ac4112accce96bf8c8df55143423cac55099fc17c0107"} Dec 06 08:15:01 crc kubenswrapper[4945]: I1206 08:15:01.154589 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" event={"ID":"e22d9737-bf91-4dd2-8262-6e43b23ff1d2","Type":"ContainerStarted","Data":"121992cdd44efdc4613325cb3d0ab7d7bfc5393de3d08019cdaf0a7e0b6845eb"} Dec 06 08:15:01 crc kubenswrapper[4945]: I1206 08:15:01.169335 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" podStartSLOduration=1.169267997 podStartE2EDuration="1.169267997s" podCreationTimestamp="2025-12-06 08:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:15:01.167961442 +0000 UTC m=+4914.622822486" watchObservedRunningTime="2025-12-06 08:15:01.169267997 +0000 UTC m=+4914.624129041" Dec 06 08:15:02 crc kubenswrapper[4945]: I1206 08:15:02.163853 4945 generic.go:334] "Generic (PLEG): container finished" podID="e22d9737-bf91-4dd2-8262-6e43b23ff1d2" containerID="d310a26a2941cadf470ac4112accce96bf8c8df55143423cac55099fc17c0107" exitCode=0 Dec 06 08:15:02 crc kubenswrapper[4945]: I1206 08:15:02.163944 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" event={"ID":"e22d9737-bf91-4dd2-8262-6e43b23ff1d2","Type":"ContainerDied","Data":"d310a26a2941cadf470ac4112accce96bf8c8df55143423cac55099fc17c0107"} Dec 06 08:15:03 crc kubenswrapper[4945]: I1206 08:15:03.501683 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:03 crc kubenswrapper[4945]: I1206 08:15:03.631964 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp8f9\" (UniqueName: \"kubernetes.io/projected/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-kube-api-access-qp8f9\") pod \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " Dec 06 08:15:03 crc kubenswrapper[4945]: I1206 08:15:03.632122 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-secret-volume\") pod \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " Dec 06 08:15:03 crc kubenswrapper[4945]: I1206 08:15:03.632178 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-config-volume\") pod \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\" (UID: \"e22d9737-bf91-4dd2-8262-6e43b23ff1d2\") " Dec 06 08:15:03 crc kubenswrapper[4945]: I1206 08:15:03.633035 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-config-volume" (OuterVolumeSpecName: "config-volume") pod "e22d9737-bf91-4dd2-8262-6e43b23ff1d2" (UID: "e22d9737-bf91-4dd2-8262-6e43b23ff1d2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:15:03 crc kubenswrapper[4945]: I1206 08:15:03.637459 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-kube-api-access-qp8f9" (OuterVolumeSpecName: "kube-api-access-qp8f9") pod "e22d9737-bf91-4dd2-8262-6e43b23ff1d2" (UID: "e22d9737-bf91-4dd2-8262-6e43b23ff1d2"). InnerVolumeSpecName "kube-api-access-qp8f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:15:03 crc kubenswrapper[4945]: I1206 08:15:03.637934 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e22d9737-bf91-4dd2-8262-6e43b23ff1d2" (UID: "e22d9737-bf91-4dd2-8262-6e43b23ff1d2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:15:03 crc kubenswrapper[4945]: I1206 08:15:03.734349 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:03 crc kubenswrapper[4945]: I1206 08:15:03.734381 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:03 crc kubenswrapper[4945]: I1206 08:15:03.734395 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp8f9\" (UniqueName: \"kubernetes.io/projected/e22d9737-bf91-4dd2-8262-6e43b23ff1d2-kube-api-access-qp8f9\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:04 crc kubenswrapper[4945]: I1206 08:15:04.182702 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" event={"ID":"e22d9737-bf91-4dd2-8262-6e43b23ff1d2","Type":"ContainerDied","Data":"121992cdd44efdc4613325cb3d0ab7d7bfc5393de3d08019cdaf0a7e0b6845eb"} Dec 06 08:15:04 crc kubenswrapper[4945]: I1206 08:15:04.182755 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j" Dec 06 08:15:04 crc kubenswrapper[4945]: I1206 08:15:04.182785 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="121992cdd44efdc4613325cb3d0ab7d7bfc5393de3d08019cdaf0a7e0b6845eb" Dec 06 08:15:04 crc kubenswrapper[4945]: I1206 08:15:04.253756 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q"] Dec 06 08:15:04 crc kubenswrapper[4945]: I1206 08:15:04.263957 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416770-g8n9q"] Dec 06 08:15:04 crc kubenswrapper[4945]: I1206 08:15:04.963875 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f287003-14b7-4397-b93f-ffc3a87db524" path="/var/lib/kubelet/pods/2f287003-14b7-4397-b93f-ffc3a87db524/volumes" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.271401 4945 scope.go:117] "RemoveContainer" containerID="3cf427c5a6ae004a4d62091b57effad841cd9e1e41fdee962577561e499dbeb5" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.284503 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 08:15:24 crc kubenswrapper[4945]: E1206 08:15:24.284895 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22d9737-bf91-4dd2-8262-6e43b23ff1d2" containerName="collect-profiles" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.284919 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22d9737-bf91-4dd2-8262-6e43b23ff1d2" containerName="collect-profiles" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.285117 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22d9737-bf91-4dd2-8262-6e43b23ff1d2" containerName="collect-profiles" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.286117 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.299802 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.300044 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-gwb5l" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.300320 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.313015 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.316854 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.328964 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.344819 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.352351 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.360453 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.372442 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.444953 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/41bd1596-3908-44e6-840c-3011459fdc81-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445124 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/62420d57-1545-4d7e-834f-60d8c6a6a4ff-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445206 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q55fx\" (UniqueName: \"kubernetes.io/projected/62420d57-1545-4d7e-834f-60d8c6a6a4ff-kube-api-access-q55fx\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445334 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5f50f1d6-54db-4e4b-8e27-958c63c2c0a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f50f1d6-54db-4e4b-8e27-958c63c2c0a9\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445394 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/befb328f-40fa-41b6-845e-79c19af5ce49-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445421 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fb41e4c0-ab57-4ab6-85a6-6e84c0e466b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb41e4c0-ab57-4ab6-85a6-6e84c0e466b4\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445449 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-39060f96-6683-4c2d-9157-49aaa9c49970\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39060f96-6683-4c2d-9157-49aaa9c49970\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445475 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kjxq\" (UniqueName: \"kubernetes.io/projected/41bd1596-3908-44e6-840c-3011459fdc81-kube-api-access-9kjxq\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445545 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/befb328f-40fa-41b6-845e-79c19af5ce49-config\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445657 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/befb328f-40fa-41b6-845e-79c19af5ce49-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445699 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bd1596-3908-44e6-840c-3011459fdc81-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445726 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41bd1596-3908-44e6-840c-3011459fdc81-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445757 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/befb328f-40fa-41b6-845e-79c19af5ce49-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445826 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fps8d\" (UniqueName: \"kubernetes.io/projected/befb328f-40fa-41b6-845e-79c19af5ce49-kube-api-access-fps8d\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445879 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62420d57-1545-4d7e-834f-60d8c6a6a4ff-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.445993 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62420d57-1545-4d7e-834f-60d8c6a6a4ff-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.446090 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62420d57-1545-4d7e-834f-60d8c6a6a4ff-config\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.446179 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41bd1596-3908-44e6-840c-3011459fdc81-config\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.502891 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.504331 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.507075 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.507235 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-dcxdw" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.507813 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.512186 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.525908 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.527380 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.531669 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.532981 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.539931 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.545926 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548617 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62420d57-1545-4d7e-834f-60d8c6a6a4ff-config\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548670 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41bd1596-3908-44e6-840c-3011459fdc81-config\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548692 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/41bd1596-3908-44e6-840c-3011459fdc81-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548716 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/62420d57-1545-4d7e-834f-60d8c6a6a4ff-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548740 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q55fx\" (UniqueName: \"kubernetes.io/projected/62420d57-1545-4d7e-834f-60d8c6a6a4ff-kube-api-access-q55fx\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548762 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5f50f1d6-54db-4e4b-8e27-958c63c2c0a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f50f1d6-54db-4e4b-8e27-958c63c2c0a9\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548782 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/befb328f-40fa-41b6-845e-79c19af5ce49-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548808 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fb41e4c0-ab57-4ab6-85a6-6e84c0e466b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb41e4c0-ab57-4ab6-85a6-6e84c0e466b4\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548832 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-39060f96-6683-4c2d-9157-49aaa9c49970\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39060f96-6683-4c2d-9157-49aaa9c49970\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548850 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kjxq\" (UniqueName: \"kubernetes.io/projected/41bd1596-3908-44e6-840c-3011459fdc81-kube-api-access-9kjxq\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548875 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/befb328f-40fa-41b6-845e-79c19af5ce49-config\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548890 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/befb328f-40fa-41b6-845e-79c19af5ce49-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548907 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bd1596-3908-44e6-840c-3011459fdc81-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548924 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41bd1596-3908-44e6-840c-3011459fdc81-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548940 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/befb328f-40fa-41b6-845e-79c19af5ce49-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548958 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fps8d\" (UniqueName: \"kubernetes.io/projected/befb328f-40fa-41b6-845e-79c19af5ce49-kube-api-access-fps8d\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.548975 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62420d57-1545-4d7e-834f-60d8c6a6a4ff-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.549009 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62420d57-1545-4d7e-834f-60d8c6a6a4ff-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.549729 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62420d57-1545-4d7e-834f-60d8c6a6a4ff-config\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.550393 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/41bd1596-3908-44e6-840c-3011459fdc81-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.550927 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/62420d57-1545-4d7e-834f-60d8c6a6a4ff-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.553686 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bd1596-3908-44e6-840c-3011459fdc81-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.554950 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62420d57-1545-4d7e-834f-60d8c6a6a4ff-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.556363 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41bd1596-3908-44e6-840c-3011459fdc81-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.557150 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.557177 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5f50f1d6-54db-4e4b-8e27-958c63c2c0a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f50f1d6-54db-4e4b-8e27-958c63c2c0a9\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bba583c6035870e0cfed4ffc656770aa50eda21128570972136e9ff82b4aa4d3/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.557215 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.557246 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-39060f96-6683-4c2d-9157-49aaa9c49970\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39060f96-6683-4c2d-9157-49aaa9c49970\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b21f70720150c6a63469cdd7a2148c9cc213f0b6c4c3a8950cdf797c420acfb3/globalmount\"" pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.557443 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/befb328f-40fa-41b6-845e-79c19af5ce49-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.559464 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/befb328f-40fa-41b6-845e-79c19af5ce49-config\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.561651 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/befb328f-40fa-41b6-845e-79c19af5ce49-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.562532 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.562561 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fb41e4c0-ab57-4ab6-85a6-6e84c0e466b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb41e4c0-ab57-4ab6-85a6-6e84c0e466b4\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/eae96b6441350ba0451d1cfacbfcff29bd113a67291066cd2655099542714ec2/globalmount\"" pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.563446 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41bd1596-3908-44e6-840c-3011459fdc81-config\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.566244 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/befb328f-40fa-41b6-845e-79c19af5ce49-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.573319 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kjxq\" (UniqueName: \"kubernetes.io/projected/41bd1596-3908-44e6-840c-3011459fdc81-kube-api-access-9kjxq\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.575583 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62420d57-1545-4d7e-834f-60d8c6a6a4ff-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.580553 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q55fx\" (UniqueName: \"kubernetes.io/projected/62420d57-1545-4d7e-834f-60d8c6a6a4ff-kube-api-access-q55fx\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.584543 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fps8d\" (UniqueName: \"kubernetes.io/projected/befb328f-40fa-41b6-845e-79c19af5ce49-kube-api-access-fps8d\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.597455 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fb41e4c0-ab57-4ab6-85a6-6e84c0e466b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb41e4c0-ab57-4ab6-85a6-6e84c0e466b4\") pod \"ovsdbserver-nb-1\" (UID: \"41bd1596-3908-44e6-840c-3011459fdc81\") " pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.597777 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-39060f96-6683-4c2d-9157-49aaa9c49970\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39060f96-6683-4c2d-9157-49aaa9c49970\") pod \"ovsdbserver-nb-2\" (UID: \"62420d57-1545-4d7e-834f-60d8c6a6a4ff\") " pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.610600 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5f50f1d6-54db-4e4b-8e27-958c63c2c0a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f50f1d6-54db-4e4b-8e27-958c63c2c0a9\") pod \"ovsdbserver-nb-0\" (UID: \"befb328f-40fa-41b6-845e-79c19af5ce49\") " pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.650938 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/744a0d22-5718-4d6a-ad18-2f522e666977-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651326 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/744a0d22-5718-4d6a-ad18-2f522e666977-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651405 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97504023-3ba7-4480-9c54-ff4985c3f38c-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651471 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97504023-3ba7-4480-9c54-ff4985c3f38c-config\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651568 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zbkp\" (UniqueName: \"kubernetes.io/projected/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-kube-api-access-8zbkp\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651620 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97504023-3ba7-4480-9c54-ff4985c3f38c-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651675 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-521b3920-1e69-4404-9adc-cac37525dffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-521b3920-1e69-4404-9adc-cac37525dffd\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651709 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/744a0d22-5718-4d6a-ad18-2f522e666977-config\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651734 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-config\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651770 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-467f13c9-d3c5-4db3-84a8-5dd1a303ebe8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-467f13c9-d3c5-4db3-84a8-5dd1a303ebe8\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651812 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651843 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/744a0d22-5718-4d6a-ad18-2f522e666977-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.651892 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ecf23374-1c66-4f8e-ad4f-653478850b5f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ecf23374-1c66-4f8e-ad4f-653478850b5f\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.652208 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.652377 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvw86\" (UniqueName: \"kubernetes.io/projected/97504023-3ba7-4480-9c54-ff4985c3f38c-kube-api-access-jvw86\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.652425 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d7dt\" (UniqueName: \"kubernetes.io/projected/744a0d22-5718-4d6a-ad18-2f522e666977-kube-api-access-7d7dt\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.652474 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.652506 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97504023-3ba7-4480-9c54-ff4985c3f38c-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.667748 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.678682 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754150 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/744a0d22-5718-4d6a-ad18-2f522e666977-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754197 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/744a0d22-5718-4d6a-ad18-2f522e666977-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97504023-3ba7-4480-9c54-ff4985c3f38c-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754262 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754304 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97504023-3ba7-4480-9c54-ff4985c3f38c-config\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754321 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zbkp\" (UniqueName: \"kubernetes.io/projected/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-kube-api-access-8zbkp\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754338 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97504023-3ba7-4480-9c54-ff4985c3f38c-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754356 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-521b3920-1e69-4404-9adc-cac37525dffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-521b3920-1e69-4404-9adc-cac37525dffd\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754382 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/744a0d22-5718-4d6a-ad18-2f522e666977-config\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754400 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-config\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754421 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-467f13c9-d3c5-4db3-84a8-5dd1a303ebe8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-467f13c9-d3c5-4db3-84a8-5dd1a303ebe8\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754436 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754450 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/744a0d22-5718-4d6a-ad18-2f522e666977-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754474 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ecf23374-1c66-4f8e-ad4f-653478850b5f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ecf23374-1c66-4f8e-ad4f-653478850b5f\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754498 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvw86\" (UniqueName: \"kubernetes.io/projected/97504023-3ba7-4480-9c54-ff4985c3f38c-kube-api-access-jvw86\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754515 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d7dt\" (UniqueName: \"kubernetes.io/projected/744a0d22-5718-4d6a-ad18-2f522e666977-kube-api-access-7d7dt\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754532 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754549 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97504023-3ba7-4480-9c54-ff4985c3f38c-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.754993 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97504023-3ba7-4480-9c54-ff4985c3f38c-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.755834 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/744a0d22-5718-4d6a-ad18-2f522e666977-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.757324 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-config\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.758099 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/744a0d22-5718-4d6a-ad18-2f522e666977-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.758776 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97504023-3ba7-4480-9c54-ff4985c3f38c-config\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.759048 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.759072 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97504023-3ba7-4480-9c54-ff4985c3f38c-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.759105 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/744a0d22-5718-4d6a-ad18-2f522e666977-config\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.763153 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.763201 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ecf23374-1c66-4f8e-ad4f-653478850b5f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ecf23374-1c66-4f8e-ad4f-653478850b5f\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/76902cc4e143b53400dd6970977b56f28e9373290ba92acdfa69da104ee92a49/globalmount\"" pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.763207 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.763250 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-467f13c9-d3c5-4db3-84a8-5dd1a303ebe8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-467f13c9-d3c5-4db3-84a8-5dd1a303ebe8\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1aaf4ffaed3a8e3c3897a54ba76008c4331b2bcdc38c8e101879ac8173f9b7ce/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.763636 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.764687 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/744a0d22-5718-4d6a-ad18-2f522e666977-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.766314 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.769357 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-521b3920-1e69-4404-9adc-cac37525dffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-521b3920-1e69-4404-9adc-cac37525dffd\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/804ed36138f85b9494e3e5cad90809c2395c32849cca6aaf2a2477c5049b0c61/globalmount\"" pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.767043 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97504023-3ba7-4480-9c54-ff4985c3f38c-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.772183 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.773836 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zbkp\" (UniqueName: \"kubernetes.io/projected/b14a3e86-54c1-46bd-b470-3ae5e648b2ff-kube-api-access-8zbkp\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.779048 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvw86\" (UniqueName: \"kubernetes.io/projected/97504023-3ba7-4480-9c54-ff4985c3f38c-kube-api-access-jvw86\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.780272 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d7dt\" (UniqueName: \"kubernetes.io/projected/744a0d22-5718-4d6a-ad18-2f522e666977-kube-api-access-7d7dt\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.806297 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ecf23374-1c66-4f8e-ad4f-653478850b5f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ecf23374-1c66-4f8e-ad4f-653478850b5f\") pod \"ovsdbserver-sb-1\" (UID: \"744a0d22-5718-4d6a-ad18-2f522e666977\") " pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.812670 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-467f13c9-d3c5-4db3-84a8-5dd1a303ebe8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-467f13c9-d3c5-4db3-84a8-5dd1a303ebe8\") pod \"ovsdbserver-sb-0\" (UID: \"b14a3e86-54c1-46bd-b470-3ae5e648b2ff\") " pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.817598 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-521b3920-1e69-4404-9adc-cac37525dffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-521b3920-1e69-4404-9adc-cac37525dffd\") pod \"ovsdbserver-sb-2\" (UID: \"97504023-3ba7-4480-9c54-ff4985c3f38c\") " pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.821140 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.847674 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:24 crc kubenswrapper[4945]: I1206 08:15:24.858327 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:25 crc kubenswrapper[4945]: I1206 08:15:25.047364 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 06 08:15:25 crc kubenswrapper[4945]: I1206 08:15:25.175213 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 08:15:25 crc kubenswrapper[4945]: W1206 08:15:25.175717 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbefb328f_40fa_41b6_845e_79c19af5ce49.slice/crio-0b9b46ed84b90236007755bdc2f2fc06a23aff6c70e42a087b8e235b6ff5a17d WatchSource:0}: Error finding container 0b9b46ed84b90236007755bdc2f2fc06a23aff6c70e42a087b8e235b6ff5a17d: Status 404 returned error can't find the container with id 0b9b46ed84b90236007755bdc2f2fc06a23aff6c70e42a087b8e235b6ff5a17d Dec 06 08:15:25 crc kubenswrapper[4945]: I1206 08:15:25.290351 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 06 08:15:25 crc kubenswrapper[4945]: W1206 08:15:25.304309 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62420d57_1545_4d7e_834f_60d8c6a6a4ff.slice/crio-3fc54736ba5ae68e149caef7b12c98c3005f3ca48319c0d735390bff58c9a332 WatchSource:0}: Error finding container 3fc54736ba5ae68e149caef7b12c98c3005f3ca48319c0d735390bff58c9a332: Status 404 returned error can't find the container with id 3fc54736ba5ae68e149caef7b12c98c3005f3ca48319c0d735390bff58c9a332 Dec 06 08:15:25 crc kubenswrapper[4945]: I1206 08:15:25.387478 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"41bd1596-3908-44e6-840c-3011459fdc81","Type":"ContainerStarted","Data":"3a7cf89d995df7e0831cdb78d33cb267f1dcaea02d40efbbfdaee2005ad72a57"} Dec 06 08:15:25 crc kubenswrapper[4945]: I1206 08:15:25.391466 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"befb328f-40fa-41b6-845e-79c19af5ce49","Type":"ContainerStarted","Data":"0b9b46ed84b90236007755bdc2f2fc06a23aff6c70e42a087b8e235b6ff5a17d"} Dec 06 08:15:25 crc kubenswrapper[4945]: I1206 08:15:25.393354 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"62420d57-1545-4d7e-834f-60d8c6a6a4ff","Type":"ContainerStarted","Data":"3fc54736ba5ae68e149caef7b12c98c3005f3ca48319c0d735390bff58c9a332"} Dec 06 08:15:25 crc kubenswrapper[4945]: I1206 08:15:25.477013 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 06 08:15:25 crc kubenswrapper[4945]: W1206 08:15:25.486935 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97504023_3ba7_4480_9c54_ff4985c3f38c.slice/crio-2857820bd0248fa42e9f0bbb744594e41c3e37a63626b5325481b0a1331f2d8c WatchSource:0}: Error finding container 2857820bd0248fa42e9f0bbb744594e41c3e37a63626b5325481b0a1331f2d8c: Status 404 returned error can't find the container with id 2857820bd0248fa42e9f0bbb744594e41c3e37a63626b5325481b0a1331f2d8c Dec 06 08:15:25 crc kubenswrapper[4945]: I1206 08:15:25.710746 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 08:15:26 crc kubenswrapper[4945]: I1206 08:15:26.150344 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 06 08:15:26 crc kubenswrapper[4945]: I1206 08:15:26.403821 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"744a0d22-5718-4d6a-ad18-2f522e666977","Type":"ContainerStarted","Data":"e51242a59365130024e340f9929b4915fda927648893df7611e737f522f4b11c"} Dec 06 08:15:26 crc kubenswrapper[4945]: I1206 08:15:26.404996 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b14a3e86-54c1-46bd-b470-3ae5e648b2ff","Type":"ContainerStarted","Data":"8f584a07dd1c1b69c4f513becea1590ca74dd903126a9aa0c93f8764c5cdbc52"} Dec 06 08:15:26 crc kubenswrapper[4945]: I1206 08:15:26.406885 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"97504023-3ba7-4480-9c54-ff4985c3f38c","Type":"ContainerStarted","Data":"2857820bd0248fa42e9f0bbb744594e41c3e37a63626b5325481b0a1331f2d8c"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.439949 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b14a3e86-54c1-46bd-b470-3ae5e648b2ff","Type":"ContainerStarted","Data":"35bc5de73676768c39eef57608c71df2630eb0a2bd2f65dad5cf936c61c833a4"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.440617 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b14a3e86-54c1-46bd-b470-3ae5e648b2ff","Type":"ContainerStarted","Data":"ea9572bbb80a73edbe6b4d0f88bfd4dfaebf5d9ff900b16cbe1540d901cd2cd2"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.442150 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"97504023-3ba7-4480-9c54-ff4985c3f38c","Type":"ContainerStarted","Data":"f9e51d620914565daa32f31cf707271b464a7d4475b3c7f0edf6540f417445c1"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.442203 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"97504023-3ba7-4480-9c54-ff4985c3f38c","Type":"ContainerStarted","Data":"546bd6ca77b839aa7e580a3601680641a9f844781aca7e7cd7e31ba893bf7b56"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.447025 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"744a0d22-5718-4d6a-ad18-2f522e666977","Type":"ContainerStarted","Data":"72a932c03c359dbe20deb6bfd73c8403d6b2c45d3671ac71c15534716b2e9c82"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.447061 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"744a0d22-5718-4d6a-ad18-2f522e666977","Type":"ContainerStarted","Data":"af244530546e79b98e9e8cafc7581751a84d6e62798ef4c789702d5727fcf7eb"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.450741 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"41bd1596-3908-44e6-840c-3011459fdc81","Type":"ContainerStarted","Data":"c2c7a142164231c7a54cda48c4bc8677b8769c9c6e46b563e26dcee052717b7f"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.450778 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"41bd1596-3908-44e6-840c-3011459fdc81","Type":"ContainerStarted","Data":"96654611e51cbd7bfa1f10ea38af7d39b0ce3b2a87b02d072eb8542e3228432d"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.455610 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"befb328f-40fa-41b6-845e-79c19af5ce49","Type":"ContainerStarted","Data":"a3057ef3772ac2b35802a2dfd47b6431f825002cf41419ace4acc7fc21bd1485"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.455654 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"befb328f-40fa-41b6-845e-79c19af5ce49","Type":"ContainerStarted","Data":"cf08fb3e46442856318d8525cc92080116455354fb8700fd298f1dbac5b9ce78"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.463333 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"62420d57-1545-4d7e-834f-60d8c6a6a4ff","Type":"ContainerStarted","Data":"108b60ea49aa77c0499d64b445285d02f1f249a0d4c927ca02d1df6454d6587e"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.463373 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"62420d57-1545-4d7e-834f-60d8c6a6a4ff","Type":"ContainerStarted","Data":"a1d710c8a0a3617f63d4e99bd5c4db02903022113ba65a901ac4f8fff44f355d"} Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.463398 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.794100972 podStartE2EDuration="7.463377481s" podCreationTimestamp="2025-12-06 08:15:23 +0000 UTC" firstStartedPulling="2025-12-06 08:15:25.724580232 +0000 UTC m=+4939.179441286" lastFinishedPulling="2025-12-06 08:15:29.393856741 +0000 UTC m=+4942.848717795" observedRunningTime="2025-12-06 08:15:30.4603308 +0000 UTC m=+4943.915191844" watchObservedRunningTime="2025-12-06 08:15:30.463377481 +0000 UTC m=+4943.918238535" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.482849 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.137133269 podStartE2EDuration="7.482830471s" podCreationTimestamp="2025-12-06 08:15:23 +0000 UTC" firstStartedPulling="2025-12-06 08:15:25.056266966 +0000 UTC m=+4938.511128010" lastFinishedPulling="2025-12-06 08:15:29.401964168 +0000 UTC m=+4942.856825212" observedRunningTime="2025-12-06 08:15:30.477393716 +0000 UTC m=+4943.932254770" watchObservedRunningTime="2025-12-06 08:15:30.482830471 +0000 UTC m=+4943.937691515" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.501642 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.27707599 podStartE2EDuration="7.501627103s" podCreationTimestamp="2025-12-06 08:15:23 +0000 UTC" firstStartedPulling="2025-12-06 08:15:25.178098363 +0000 UTC m=+4938.632959407" lastFinishedPulling="2025-12-06 08:15:29.402649476 +0000 UTC m=+4942.857510520" observedRunningTime="2025-12-06 08:15:30.497835322 +0000 UTC m=+4943.952696366" watchObservedRunningTime="2025-12-06 08:15:30.501627103 +0000 UTC m=+4943.956488147" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.515595 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.626520862 podStartE2EDuration="7.515574346s" podCreationTimestamp="2025-12-06 08:15:23 +0000 UTC" firstStartedPulling="2025-12-06 08:15:25.492169109 +0000 UTC m=+4938.947030153" lastFinishedPulling="2025-12-06 08:15:29.381222593 +0000 UTC m=+4942.836083637" observedRunningTime="2025-12-06 08:15:30.511462556 +0000 UTC m=+4943.966323600" watchObservedRunningTime="2025-12-06 08:15:30.515574346 +0000 UTC m=+4943.970435390" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.535157 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.154811605 podStartE2EDuration="7.535133229s" podCreationTimestamp="2025-12-06 08:15:23 +0000 UTC" firstStartedPulling="2025-12-06 08:15:26.159426757 +0000 UTC m=+4939.614287801" lastFinishedPulling="2025-12-06 08:15:29.539748381 +0000 UTC m=+4942.994609425" observedRunningTime="2025-12-06 08:15:30.528843601 +0000 UTC m=+4943.983704645" watchObservedRunningTime="2025-12-06 08:15:30.535133229 +0000 UTC m=+4943.989994273" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.544424 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.415628525 podStartE2EDuration="7.544406767s" podCreationTimestamp="2025-12-06 08:15:23 +0000 UTC" firstStartedPulling="2025-12-06 08:15:25.306757193 +0000 UTC m=+4938.761618237" lastFinishedPulling="2025-12-06 08:15:29.435535435 +0000 UTC m=+4942.890396479" observedRunningTime="2025-12-06 08:15:30.54302403 +0000 UTC m=+4943.997885074" watchObservedRunningTime="2025-12-06 08:15:30.544406767 +0000 UTC m=+4943.999267811" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.653186 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.669351 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.679680 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.822377 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.859548 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:30 crc kubenswrapper[4945]: I1206 08:15:30.859692 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.708262 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.709012 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.712992 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.713372 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.737041 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.737404 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.873912 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.876555 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.897985 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.898484 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.898652 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:33 crc kubenswrapper[4945]: I1206 08:15:33.899010 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.542791 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.543425 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.544143 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.545623 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.545684 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.548605 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.772587 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-766574c965-wm27x"] Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.781403 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.788013 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.808251 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766574c965-wm27x"] Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.831529 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-ovsdbserver-sb\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.831575 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-config\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.831600 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-dns-svc\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.831855 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gxmj\" (UniqueName: \"kubernetes.io/projected/f32270e5-a81a-421e-8614-ee943449c71c-kube-api-access-8gxmj\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.934998 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gxmj\" (UniqueName: \"kubernetes.io/projected/f32270e5-a81a-421e-8614-ee943449c71c-kube-api-access-8gxmj\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.935134 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-ovsdbserver-sb\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.935161 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-config\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.935184 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-dns-svc\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.938211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-config\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.939066 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-dns-svc\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.939526 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-ovsdbserver-sb\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:34 crc kubenswrapper[4945]: I1206 08:15:34.956719 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gxmj\" (UniqueName: \"kubernetes.io/projected/f32270e5-a81a-421e-8614-ee943449c71c-kube-api-access-8gxmj\") pod \"dnsmasq-dns-766574c965-wm27x\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.020881 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766574c965-wm27x"] Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.021773 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.051953 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65bf94dc4c-szkv6"] Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.060539 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.062088 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.075600 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65bf94dc4c-szkv6"] Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.138251 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-config\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.138357 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-dns-svc\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.138390 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-nb\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.138453 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fplc\" (UniqueName: \"kubernetes.io/projected/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-kube-api-access-4fplc\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.138664 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-sb\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.246199 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-config\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.246605 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-dns-svc\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.246641 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-nb\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.246709 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fplc\" (UniqueName: \"kubernetes.io/projected/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-kube-api-access-4fplc\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.246774 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-sb\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.247764 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-sb\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.248555 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-config\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.249044 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-nb\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.253708 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-dns-svc\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.274181 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fplc\" (UniqueName: \"kubernetes.io/projected/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-kube-api-access-4fplc\") pod \"dnsmasq-dns-65bf94dc4c-szkv6\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.346119 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766574c965-wm27x"] Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.428687 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.515222 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766574c965-wm27x" event={"ID":"f32270e5-a81a-421e-8614-ee943449c71c","Type":"ContainerStarted","Data":"90a32f5557d24186cc0aac968473b5d40062ef75fdfee70c96817ff63d8a5524"} Dec 06 08:15:35 crc kubenswrapper[4945]: I1206 08:15:35.916755 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65bf94dc4c-szkv6"] Dec 06 08:15:35 crc kubenswrapper[4945]: W1206 08:15:35.920655 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod526d2c6c_3a06_44e8_8def_4ef3fff34ac9.slice/crio-f9ced3d73ca847d5588a939e670beb23000be3761e294756649c94db6d23c772 WatchSource:0}: Error finding container f9ced3d73ca847d5588a939e670beb23000be3761e294756649c94db6d23c772: Status 404 returned error can't find the container with id f9ced3d73ca847d5588a939e670beb23000be3761e294756649c94db6d23c772 Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.522043 4945 generic.go:334] "Generic (PLEG): container finished" podID="526d2c6c-3a06-44e8-8def-4ef3fff34ac9" containerID="48c040a1e3360d33634d04e1fd88cddb12860a2ce18399f08577c03561062694" exitCode=0 Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.522099 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" event={"ID":"526d2c6c-3a06-44e8-8def-4ef3fff34ac9","Type":"ContainerDied","Data":"48c040a1e3360d33634d04e1fd88cddb12860a2ce18399f08577c03561062694"} Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.522447 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" event={"ID":"526d2c6c-3a06-44e8-8def-4ef3fff34ac9","Type":"ContainerStarted","Data":"f9ced3d73ca847d5588a939e670beb23000be3761e294756649c94db6d23c772"} Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.523847 4945 generic.go:334] "Generic (PLEG): container finished" podID="f32270e5-a81a-421e-8614-ee943449c71c" containerID="f2e20c852aa8b27a9a7db22d2120ac3844ba235495db3e08a563681b59819a82" exitCode=0 Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.523877 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766574c965-wm27x" event={"ID":"f32270e5-a81a-421e-8614-ee943449c71c","Type":"ContainerDied","Data":"f2e20c852aa8b27a9a7db22d2120ac3844ba235495db3e08a563681b59819a82"} Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.793822 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.871255 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-config\") pod \"f32270e5-a81a-421e-8614-ee943449c71c\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.871470 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gxmj\" (UniqueName: \"kubernetes.io/projected/f32270e5-a81a-421e-8614-ee943449c71c-kube-api-access-8gxmj\") pod \"f32270e5-a81a-421e-8614-ee943449c71c\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.871521 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-ovsdbserver-sb\") pod \"f32270e5-a81a-421e-8614-ee943449c71c\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.871567 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-dns-svc\") pod \"f32270e5-a81a-421e-8614-ee943449c71c\" (UID: \"f32270e5-a81a-421e-8614-ee943449c71c\") " Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.875011 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32270e5-a81a-421e-8614-ee943449c71c-kube-api-access-8gxmj" (OuterVolumeSpecName: "kube-api-access-8gxmj") pod "f32270e5-a81a-421e-8614-ee943449c71c" (UID: "f32270e5-a81a-421e-8614-ee943449c71c"). InnerVolumeSpecName "kube-api-access-8gxmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.891729 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f32270e5-a81a-421e-8614-ee943449c71c" (UID: "f32270e5-a81a-421e-8614-ee943449c71c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.892469 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f32270e5-a81a-421e-8614-ee943449c71c" (UID: "f32270e5-a81a-421e-8614-ee943449c71c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.899091 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-config" (OuterVolumeSpecName: "config") pod "f32270e5-a81a-421e-8614-ee943449c71c" (UID: "f32270e5-a81a-421e-8614-ee943449c71c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.972924 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gxmj\" (UniqueName: \"kubernetes.io/projected/f32270e5-a81a-421e-8614-ee943449c71c-kube-api-access-8gxmj\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.972962 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.972975 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:36 crc kubenswrapper[4945]: I1206 08:15:36.972986 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32270e5-a81a-421e-8614-ee943449c71c-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.540825 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" event={"ID":"526d2c6c-3a06-44e8-8def-4ef3fff34ac9","Type":"ContainerStarted","Data":"c9de46e2658c6071dc8376f6b0acc2d5183c3f17f8b3bd5b7a5efc36d51e8aba"} Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.543961 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.544981 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766574c965-wm27x" Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.544667 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766574c965-wm27x" event={"ID":"f32270e5-a81a-421e-8614-ee943449c71c","Type":"ContainerDied","Data":"90a32f5557d24186cc0aac968473b5d40062ef75fdfee70c96817ff63d8a5524"} Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.546431 4945 scope.go:117] "RemoveContainer" containerID="f2e20c852aa8b27a9a7db22d2120ac3844ba235495db3e08a563681b59819a82" Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.561998 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" podStartSLOduration=2.561978005 podStartE2EDuration="2.561978005s" podCreationTimestamp="2025-12-06 08:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:15:37.560790603 +0000 UTC m=+4951.015651667" watchObservedRunningTime="2025-12-06 08:15:37.561978005 +0000 UTC m=+4951.016839049" Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.611956 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766574c965-wm27x"] Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.617749 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-766574c965-wm27x"] Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.985706 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Dec 06 08:15:37 crc kubenswrapper[4945]: E1206 08:15:37.986168 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32270e5-a81a-421e-8614-ee943449c71c" containerName="init" Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.986182 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32270e5-a81a-421e-8614-ee943449c71c" containerName="init" Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.986448 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f32270e5-a81a-421e-8614-ee943449c71c" containerName="init" Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.987267 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 06 08:15:37 crc kubenswrapper[4945]: I1206 08:15:37.989647 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.042807 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.094469 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.094523 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\") pod \"ovn-copy-data\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.094609 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkd85\" (UniqueName: \"kubernetes.io/projected/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-kube-api-access-nkd85\") pod \"ovn-copy-data\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.196603 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.196681 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\") pod \"ovn-copy-data\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.196747 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkd85\" (UniqueName: \"kubernetes.io/projected/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-kube-api-access-nkd85\") pod \"ovn-copy-data\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.200420 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.200461 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\") pod \"ovn-copy-data\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a33944c9fc7bd176a2b3f9e95e55b306f74b7493c96490098f1b95d4854f6c1d/globalmount\"" pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.203297 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.214482 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkd85\" (UniqueName: \"kubernetes.io/projected/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-kube-api-access-nkd85\") pod \"ovn-copy-data\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.265581 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\") pod \"ovn-copy-data\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.356623 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.872854 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 06 08:15:38 crc kubenswrapper[4945]: W1206 08:15:38.877514 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cad8b6b_afb4_407a_8bbd_b3e407fb957a.slice/crio-3182645f89f4ac6300b6379cd7cb04dd80a7eb1f86fc30559ee0cba5d6e7b491 WatchSource:0}: Error finding container 3182645f89f4ac6300b6379cd7cb04dd80a7eb1f86fc30559ee0cba5d6e7b491: Status 404 returned error can't find the container with id 3182645f89f4ac6300b6379cd7cb04dd80a7eb1f86fc30559ee0cba5d6e7b491 Dec 06 08:15:38 crc kubenswrapper[4945]: I1206 08:15:38.961787 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f32270e5-a81a-421e-8614-ee943449c71c" path="/var/lib/kubelet/pods/f32270e5-a81a-421e-8614-ee943449c71c/volumes" Dec 06 08:15:39 crc kubenswrapper[4945]: I1206 08:15:39.562296 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3cad8b6b-afb4-407a-8bbd-b3e407fb957a","Type":"ContainerStarted","Data":"8eb065542828012f6bc66c86d09875029a70ae45331820fc410c77a44d997877"} Dec 06 08:15:39 crc kubenswrapper[4945]: I1206 08:15:39.562661 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3cad8b6b-afb4-407a-8bbd-b3e407fb957a","Type":"ContainerStarted","Data":"3182645f89f4ac6300b6379cd7cb04dd80a7eb1f86fc30559ee0cba5d6e7b491"} Dec 06 08:15:39 crc kubenswrapper[4945]: I1206 08:15:39.574532 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.34167372 podStartE2EDuration="3.574513035s" podCreationTimestamp="2025-12-06 08:15:36 +0000 UTC" firstStartedPulling="2025-12-06 08:15:38.880074781 +0000 UTC m=+4952.334935825" lastFinishedPulling="2025-12-06 08:15:39.112914096 +0000 UTC m=+4952.567775140" observedRunningTime="2025-12-06 08:15:39.574333191 +0000 UTC m=+4953.029194255" watchObservedRunningTime="2025-12-06 08:15:39.574513035 +0000 UTC m=+4953.029374089" Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.431578 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.569017 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f7f6bbcbf-26ngg"] Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.569299 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" podUID="30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" containerName="dnsmasq-dns" containerID="cri-o://7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36" gracePeriod=10 Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.821455 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.823477 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.830973 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-l6wkp" Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.834367 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.834813 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.835928 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.923746 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ba9377fd-f829-45bb-8478-a233f5cb6a81-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.923799 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba9377fd-f829-45bb-8478-a233f5cb6a81-scripts\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.923832 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba9377fd-f829-45bb-8478-a233f5cb6a81-config\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.923880 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls9jh\" (UniqueName: \"kubernetes.io/projected/ba9377fd-f829-45bb-8478-a233f5cb6a81-kube-api-access-ls9jh\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:45 crc kubenswrapper[4945]: I1206 08:15:45.923911 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba9377fd-f829-45bb-8478-a233f5cb6a81-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.027306 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ba9377fd-f829-45bb-8478-a233f5cb6a81-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.027359 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba9377fd-f829-45bb-8478-a233f5cb6a81-scripts\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.027392 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba9377fd-f829-45bb-8478-a233f5cb6a81-config\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.027427 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls9jh\" (UniqueName: \"kubernetes.io/projected/ba9377fd-f829-45bb-8478-a233f5cb6a81-kube-api-access-ls9jh\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.027460 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba9377fd-f829-45bb-8478-a233f5cb6a81-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.033653 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba9377fd-f829-45bb-8478-a233f5cb6a81-config\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.033840 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba9377fd-f829-45bb-8478-a233f5cb6a81-scripts\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.034158 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ba9377fd-f829-45bb-8478-a233f5cb6a81-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.034886 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba9377fd-f829-45bb-8478-a233f5cb6a81-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.052826 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls9jh\" (UniqueName: \"kubernetes.io/projected/ba9377fd-f829-45bb-8478-a233f5cb6a81-kube-api-access-ls9jh\") pod \"ovn-northd-0\" (UID: \"ba9377fd-f829-45bb-8478-a233f5cb6a81\") " pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.096802 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.128519 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-dns-svc\") pod \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.128655 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xld47\" (UniqueName: \"kubernetes.io/projected/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-kube-api-access-xld47\") pod \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.128706 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-config\") pod \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\" (UID: \"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41\") " Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.137139 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-kube-api-access-xld47" (OuterVolumeSpecName: "kube-api-access-xld47") pod "30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" (UID: "30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41"). InnerVolumeSpecName "kube-api-access-xld47". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.156817 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.186465 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-config" (OuterVolumeSpecName: "config") pod "30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" (UID: "30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.202629 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" (UID: "30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.230965 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xld47\" (UniqueName: \"kubernetes.io/projected/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-kube-api-access-xld47\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.231001 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.231010 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:46 crc kubenswrapper[4945]: W1206 08:15:46.565116 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba9377fd_f829_45bb_8478_a233f5cb6a81.slice/crio-a160bcc67461f3c6be52d35d37618f9cb565fa40ed1516a7818b0196c27c546b WatchSource:0}: Error finding container a160bcc67461f3c6be52d35d37618f9cb565fa40ed1516a7818b0196c27c546b: Status 404 returned error can't find the container with id a160bcc67461f3c6be52d35d37618f9cb565fa40ed1516a7818b0196c27c546b Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.567241 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.623977 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ba9377fd-f829-45bb-8478-a233f5cb6a81","Type":"ContainerStarted","Data":"a160bcc67461f3c6be52d35d37618f9cb565fa40ed1516a7818b0196c27c546b"} Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.625938 4945 generic.go:334] "Generic (PLEG): container finished" podID="30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" containerID="7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36" exitCode=0 Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.625972 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" event={"ID":"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41","Type":"ContainerDied","Data":"7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36"} Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.625997 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" event={"ID":"30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41","Type":"ContainerDied","Data":"e78a3741bb35400b2e1d700e10733f4efcead36b3bcb7b9a4331a3d906ddfe55"} Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.626032 4945 scope.go:117] "RemoveContainer" containerID="7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.626041 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f7f6bbcbf-26ngg" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.660460 4945 scope.go:117] "RemoveContainer" containerID="3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.673603 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f7f6bbcbf-26ngg"] Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.678869 4945 scope.go:117] "RemoveContainer" containerID="7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36" Dec 06 08:15:46 crc kubenswrapper[4945]: E1206 08:15:46.679443 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36\": container with ID starting with 7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36 not found: ID does not exist" containerID="7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.679477 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36"} err="failed to get container status \"7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36\": rpc error: code = NotFound desc = could not find container \"7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36\": container with ID starting with 7b4fb82dbfd6b0002b9c74d95c51ad90dc4bef5ce5fec3bea73d4d65abc16b36 not found: ID does not exist" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.679504 4945 scope.go:117] "RemoveContainer" containerID="3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262" Dec 06 08:15:46 crc kubenswrapper[4945]: E1206 08:15:46.679853 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262\": container with ID starting with 3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262 not found: ID does not exist" containerID="3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.679885 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262"} err="failed to get container status \"3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262\": rpc error: code = NotFound desc = could not find container \"3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262\": container with ID starting with 3b4859d87eb97eb18bc1fa4e7f544b37f3a9c1d0c3772eec1a5be8937d657262 not found: ID does not exist" Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.682055 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f7f6bbcbf-26ngg"] Dec 06 08:15:46 crc kubenswrapper[4945]: I1206 08:15:46.963623 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" path="/var/lib/kubelet/pods/30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41/volumes" Dec 06 08:15:47 crc kubenswrapper[4945]: I1206 08:15:47.666338 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ba9377fd-f829-45bb-8478-a233f5cb6a81","Type":"ContainerStarted","Data":"17f8857f882f709ad326af47014ab5c5bb0855715bd249462cbf6bbf13b69832"} Dec 06 08:15:49 crc kubenswrapper[4945]: I1206 08:15:49.346399 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ba9377fd-f829-45bb-8478-a233f5cb6a81","Type":"ContainerStarted","Data":"42fcea35c06089fa3973ad0f606e4290ae63f59d91ddd3432629254c01dab0ae"} Dec 06 08:15:49 crc kubenswrapper[4945]: I1206 08:15:49.346879 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 06 08:15:49 crc kubenswrapper[4945]: I1206 08:15:49.395692 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.618319728 podStartE2EDuration="4.395675798s" podCreationTimestamp="2025-12-06 08:15:45 +0000 UTC" firstStartedPulling="2025-12-06 08:15:46.567132615 +0000 UTC m=+4960.021993659" lastFinishedPulling="2025-12-06 08:15:47.344488685 +0000 UTC m=+4960.799349729" observedRunningTime="2025-12-06 08:15:49.387837549 +0000 UTC m=+4962.842698593" watchObservedRunningTime="2025-12-06 08:15:49.395675798 +0000 UTC m=+4962.850536842" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.339321 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-87vk8"] Dec 06 08:15:51 crc kubenswrapper[4945]: E1206 08:15:51.339953 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" containerName="dnsmasq-dns" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.339966 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" containerName="dnsmasq-dns" Dec 06 08:15:51 crc kubenswrapper[4945]: E1206 08:15:51.339992 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" containerName="init" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.339998 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" containerName="init" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.340174 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="30b41eee-ba2d-45a7-8c7b-9c8a9fe1fd41" containerName="dnsmasq-dns" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.340768 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-87vk8" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.349488 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-87vk8"] Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.407514 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b486576-3473-4507-babe-9cadb07f9d17-operator-scripts\") pod \"keystone-db-create-87vk8\" (UID: \"8b486576-3473-4507-babe-9cadb07f9d17\") " pod="openstack/keystone-db-create-87vk8" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.407596 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j79rz\" (UniqueName: \"kubernetes.io/projected/8b486576-3473-4507-babe-9cadb07f9d17-kube-api-access-j79rz\") pod \"keystone-db-create-87vk8\" (UID: \"8b486576-3473-4507-babe-9cadb07f9d17\") " pod="openstack/keystone-db-create-87vk8" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.445410 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-3f05-account-create-update-5hrms"] Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.446706 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3f05-account-create-update-5hrms" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.448914 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.453332 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3f05-account-create-update-5hrms"] Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.508824 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b486576-3473-4507-babe-9cadb07f9d17-operator-scripts\") pod \"keystone-db-create-87vk8\" (UID: \"8b486576-3473-4507-babe-9cadb07f9d17\") " pod="openstack/keystone-db-create-87vk8" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.508920 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j79rz\" (UniqueName: \"kubernetes.io/projected/8b486576-3473-4507-babe-9cadb07f9d17-kube-api-access-j79rz\") pod \"keystone-db-create-87vk8\" (UID: \"8b486576-3473-4507-babe-9cadb07f9d17\") " pod="openstack/keystone-db-create-87vk8" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.509591 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b486576-3473-4507-babe-9cadb07f9d17-operator-scripts\") pod \"keystone-db-create-87vk8\" (UID: \"8b486576-3473-4507-babe-9cadb07f9d17\") " pod="openstack/keystone-db-create-87vk8" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.528903 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j79rz\" (UniqueName: \"kubernetes.io/projected/8b486576-3473-4507-babe-9cadb07f9d17-kube-api-access-j79rz\") pod \"keystone-db-create-87vk8\" (UID: \"8b486576-3473-4507-babe-9cadb07f9d17\") " pod="openstack/keystone-db-create-87vk8" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.610496 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5hpz\" (UniqueName: \"kubernetes.io/projected/e4835330-f95a-4b6b-af02-5e780b5e35be-kube-api-access-r5hpz\") pod \"keystone-3f05-account-create-update-5hrms\" (UID: \"e4835330-f95a-4b6b-af02-5e780b5e35be\") " pod="openstack/keystone-3f05-account-create-update-5hrms" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.610602 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4835330-f95a-4b6b-af02-5e780b5e35be-operator-scripts\") pod \"keystone-3f05-account-create-update-5hrms\" (UID: \"e4835330-f95a-4b6b-af02-5e780b5e35be\") " pod="openstack/keystone-3f05-account-create-update-5hrms" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.664894 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-87vk8" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.712230 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4835330-f95a-4b6b-af02-5e780b5e35be-operator-scripts\") pod \"keystone-3f05-account-create-update-5hrms\" (UID: \"e4835330-f95a-4b6b-af02-5e780b5e35be\") " pod="openstack/keystone-3f05-account-create-update-5hrms" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.713630 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5hpz\" (UniqueName: \"kubernetes.io/projected/e4835330-f95a-4b6b-af02-5e780b5e35be-kube-api-access-r5hpz\") pod \"keystone-3f05-account-create-update-5hrms\" (UID: \"e4835330-f95a-4b6b-af02-5e780b5e35be\") " pod="openstack/keystone-3f05-account-create-update-5hrms" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.714091 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4835330-f95a-4b6b-af02-5e780b5e35be-operator-scripts\") pod \"keystone-3f05-account-create-update-5hrms\" (UID: \"e4835330-f95a-4b6b-af02-5e780b5e35be\") " pod="openstack/keystone-3f05-account-create-update-5hrms" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.733930 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5hpz\" (UniqueName: \"kubernetes.io/projected/e4835330-f95a-4b6b-af02-5e780b5e35be-kube-api-access-r5hpz\") pod \"keystone-3f05-account-create-update-5hrms\" (UID: \"e4835330-f95a-4b6b-af02-5e780b5e35be\") " pod="openstack/keystone-3f05-account-create-update-5hrms" Dec 06 08:15:51 crc kubenswrapper[4945]: I1206 08:15:51.765837 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3f05-account-create-update-5hrms" Dec 06 08:15:52 crc kubenswrapper[4945]: W1206 08:15:52.085680 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b486576_3473_4507_babe_9cadb07f9d17.slice/crio-5a16729a95ffee0db2951df118d97ca996bd7e38b285e2ef47d13a8f4164c357 WatchSource:0}: Error finding container 5a16729a95ffee0db2951df118d97ca996bd7e38b285e2ef47d13a8f4164c357: Status 404 returned error can't find the container with id 5a16729a95ffee0db2951df118d97ca996bd7e38b285e2ef47d13a8f4164c357 Dec 06 08:15:52 crc kubenswrapper[4945]: I1206 08:15:52.086657 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-87vk8"] Dec 06 08:15:52 crc kubenswrapper[4945]: I1206 08:15:52.218264 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3f05-account-create-update-5hrms"] Dec 06 08:15:52 crc kubenswrapper[4945]: I1206 08:15:52.369698 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-87vk8" event={"ID":"8b486576-3473-4507-babe-9cadb07f9d17","Type":"ContainerStarted","Data":"6cb22d807b8bd6a4e00c6de4998d605c92b6fabfe0dc506cee6463e3ce354e41"} Dec 06 08:15:52 crc kubenswrapper[4945]: I1206 08:15:52.369737 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-87vk8" event={"ID":"8b486576-3473-4507-babe-9cadb07f9d17","Type":"ContainerStarted","Data":"5a16729a95ffee0db2951df118d97ca996bd7e38b285e2ef47d13a8f4164c357"} Dec 06 08:15:52 crc kubenswrapper[4945]: I1206 08:15:52.371661 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3f05-account-create-update-5hrms" event={"ID":"e4835330-f95a-4b6b-af02-5e780b5e35be","Type":"ContainerStarted","Data":"5d7d7fff4ebe622d4e57e30a216d7afec2b897cc2b259598d1d2f4f77e39bbe0"} Dec 06 08:15:52 crc kubenswrapper[4945]: I1206 08:15:52.371714 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3f05-account-create-update-5hrms" event={"ID":"e4835330-f95a-4b6b-af02-5e780b5e35be","Type":"ContainerStarted","Data":"740fdbcce207dd6839f73ccc70c1aae6fbb93143e64f913420e50cda55769c4f"} Dec 06 08:15:52 crc kubenswrapper[4945]: I1206 08:15:52.387781 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-87vk8" podStartSLOduration=1.3877645539999999 podStartE2EDuration="1.387764554s" podCreationTimestamp="2025-12-06 08:15:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:15:52.386799098 +0000 UTC m=+4965.841660142" watchObservedRunningTime="2025-12-06 08:15:52.387764554 +0000 UTC m=+4965.842625598" Dec 06 08:15:52 crc kubenswrapper[4945]: I1206 08:15:52.411684 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-3f05-account-create-update-5hrms" podStartSLOduration=1.411661853 podStartE2EDuration="1.411661853s" podCreationTimestamp="2025-12-06 08:15:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:15:52.407194543 +0000 UTC m=+4965.862055607" watchObservedRunningTime="2025-12-06 08:15:52.411661853 +0000 UTC m=+4965.866522897" Dec 06 08:15:53 crc kubenswrapper[4945]: I1206 08:15:53.381479 4945 generic.go:334] "Generic (PLEG): container finished" podID="e4835330-f95a-4b6b-af02-5e780b5e35be" containerID="5d7d7fff4ebe622d4e57e30a216d7afec2b897cc2b259598d1d2f4f77e39bbe0" exitCode=0 Dec 06 08:15:53 crc kubenswrapper[4945]: I1206 08:15:53.381560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3f05-account-create-update-5hrms" event={"ID":"e4835330-f95a-4b6b-af02-5e780b5e35be","Type":"ContainerDied","Data":"5d7d7fff4ebe622d4e57e30a216d7afec2b897cc2b259598d1d2f4f77e39bbe0"} Dec 06 08:15:53 crc kubenswrapper[4945]: I1206 08:15:53.383358 4945 generic.go:334] "Generic (PLEG): container finished" podID="8b486576-3473-4507-babe-9cadb07f9d17" containerID="6cb22d807b8bd6a4e00c6de4998d605c92b6fabfe0dc506cee6463e3ce354e41" exitCode=0 Dec 06 08:15:53 crc kubenswrapper[4945]: I1206 08:15:53.383409 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-87vk8" event={"ID":"8b486576-3473-4507-babe-9cadb07f9d17","Type":"ContainerDied","Data":"6cb22d807b8bd6a4e00c6de4998d605c92b6fabfe0dc506cee6463e3ce354e41"} Dec 06 08:15:54 crc kubenswrapper[4945]: I1206 08:15:54.846114 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-87vk8" Dec 06 08:15:54 crc kubenswrapper[4945]: I1206 08:15:54.852073 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3f05-account-create-update-5hrms" Dec 06 08:15:54 crc kubenswrapper[4945]: I1206 08:15:54.964479 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5hpz\" (UniqueName: \"kubernetes.io/projected/e4835330-f95a-4b6b-af02-5e780b5e35be-kube-api-access-r5hpz\") pod \"e4835330-f95a-4b6b-af02-5e780b5e35be\" (UID: \"e4835330-f95a-4b6b-af02-5e780b5e35be\") " Dec 06 08:15:54 crc kubenswrapper[4945]: I1206 08:15:54.964672 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4835330-f95a-4b6b-af02-5e780b5e35be-operator-scripts\") pod \"e4835330-f95a-4b6b-af02-5e780b5e35be\" (UID: \"e4835330-f95a-4b6b-af02-5e780b5e35be\") " Dec 06 08:15:54 crc kubenswrapper[4945]: I1206 08:15:54.964726 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b486576-3473-4507-babe-9cadb07f9d17-operator-scripts\") pod \"8b486576-3473-4507-babe-9cadb07f9d17\" (UID: \"8b486576-3473-4507-babe-9cadb07f9d17\") " Dec 06 08:15:54 crc kubenswrapper[4945]: I1206 08:15:54.964746 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j79rz\" (UniqueName: \"kubernetes.io/projected/8b486576-3473-4507-babe-9cadb07f9d17-kube-api-access-j79rz\") pod \"8b486576-3473-4507-babe-9cadb07f9d17\" (UID: \"8b486576-3473-4507-babe-9cadb07f9d17\") " Dec 06 08:15:54 crc kubenswrapper[4945]: I1206 08:15:54.965900 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4835330-f95a-4b6b-af02-5e780b5e35be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e4835330-f95a-4b6b-af02-5e780b5e35be" (UID: "e4835330-f95a-4b6b-af02-5e780b5e35be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:15:54 crc kubenswrapper[4945]: I1206 08:15:54.966534 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b486576-3473-4507-babe-9cadb07f9d17-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b486576-3473-4507-babe-9cadb07f9d17" (UID: "8b486576-3473-4507-babe-9cadb07f9d17"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:15:54 crc kubenswrapper[4945]: I1206 08:15:54.970674 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b486576-3473-4507-babe-9cadb07f9d17-kube-api-access-j79rz" (OuterVolumeSpecName: "kube-api-access-j79rz") pod "8b486576-3473-4507-babe-9cadb07f9d17" (UID: "8b486576-3473-4507-babe-9cadb07f9d17"). InnerVolumeSpecName "kube-api-access-j79rz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:15:54 crc kubenswrapper[4945]: I1206 08:15:54.970864 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4835330-f95a-4b6b-af02-5e780b5e35be-kube-api-access-r5hpz" (OuterVolumeSpecName: "kube-api-access-r5hpz") pod "e4835330-f95a-4b6b-af02-5e780b5e35be" (UID: "e4835330-f95a-4b6b-af02-5e780b5e35be"). InnerVolumeSpecName "kube-api-access-r5hpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:15:55 crc kubenswrapper[4945]: I1206 08:15:55.066359 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5hpz\" (UniqueName: \"kubernetes.io/projected/e4835330-f95a-4b6b-af02-5e780b5e35be-kube-api-access-r5hpz\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:55 crc kubenswrapper[4945]: I1206 08:15:55.066402 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4835330-f95a-4b6b-af02-5e780b5e35be-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:55 crc kubenswrapper[4945]: I1206 08:15:55.066416 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b486576-3473-4507-babe-9cadb07f9d17-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:55 crc kubenswrapper[4945]: I1206 08:15:55.066429 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j79rz\" (UniqueName: \"kubernetes.io/projected/8b486576-3473-4507-babe-9cadb07f9d17-kube-api-access-j79rz\") on node \"crc\" DevicePath \"\"" Dec 06 08:15:55 crc kubenswrapper[4945]: I1206 08:15:55.398078 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3f05-account-create-update-5hrms" event={"ID":"e4835330-f95a-4b6b-af02-5e780b5e35be","Type":"ContainerDied","Data":"740fdbcce207dd6839f73ccc70c1aae6fbb93143e64f913420e50cda55769c4f"} Dec 06 08:15:55 crc kubenswrapper[4945]: I1206 08:15:55.398135 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="740fdbcce207dd6839f73ccc70c1aae6fbb93143e64f913420e50cda55769c4f" Dec 06 08:15:55 crc kubenswrapper[4945]: I1206 08:15:55.398534 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3f05-account-create-update-5hrms" Dec 06 08:15:55 crc kubenswrapper[4945]: I1206 08:15:55.400034 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-87vk8" event={"ID":"8b486576-3473-4507-babe-9cadb07f9d17","Type":"ContainerDied","Data":"5a16729a95ffee0db2951df118d97ca996bd7e38b285e2ef47d13a8f4164c357"} Dec 06 08:15:55 crc kubenswrapper[4945]: I1206 08:15:55.400073 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-87vk8" Dec 06 08:15:55 crc kubenswrapper[4945]: I1206 08:15:55.400090 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a16729a95ffee0db2951df118d97ca996bd7e38b285e2ef47d13a8f4164c357" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.891766 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-s85wb"] Dec 06 08:15:56 crc kubenswrapper[4945]: E1206 08:15:56.892550 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4835330-f95a-4b6b-af02-5e780b5e35be" containerName="mariadb-account-create-update" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.892567 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4835330-f95a-4b6b-af02-5e780b5e35be" containerName="mariadb-account-create-update" Dec 06 08:15:56 crc kubenswrapper[4945]: E1206 08:15:56.892577 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b486576-3473-4507-babe-9cadb07f9d17" containerName="mariadb-database-create" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.892583 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b486576-3473-4507-babe-9cadb07f9d17" containerName="mariadb-database-create" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.892725 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4835330-f95a-4b6b-af02-5e780b5e35be" containerName="mariadb-account-create-update" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.892753 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b486576-3473-4507-babe-9cadb07f9d17" containerName="mariadb-database-create" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.893299 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.896379 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.896591 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.896697 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r6cjp" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.896758 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.899519 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-s85wb"] Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.997096 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-combined-ca-bundle\") pod \"keystone-db-sync-s85wb\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.997376 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-config-data\") pod \"keystone-db-sync-s85wb\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:56 crc kubenswrapper[4945]: I1206 08:15:56.997628 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt28c\" (UniqueName: \"kubernetes.io/projected/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-kube-api-access-mt28c\") pod \"keystone-db-sync-s85wb\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:57 crc kubenswrapper[4945]: I1206 08:15:57.099450 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt28c\" (UniqueName: \"kubernetes.io/projected/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-kube-api-access-mt28c\") pod \"keystone-db-sync-s85wb\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:57 crc kubenswrapper[4945]: I1206 08:15:57.099557 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-combined-ca-bundle\") pod \"keystone-db-sync-s85wb\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:57 crc kubenswrapper[4945]: I1206 08:15:57.099605 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-config-data\") pod \"keystone-db-sync-s85wb\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:57 crc kubenswrapper[4945]: I1206 08:15:57.104521 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-combined-ca-bundle\") pod \"keystone-db-sync-s85wb\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:57 crc kubenswrapper[4945]: I1206 08:15:57.111511 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-config-data\") pod \"keystone-db-sync-s85wb\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:57 crc kubenswrapper[4945]: I1206 08:15:57.117435 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt28c\" (UniqueName: \"kubernetes.io/projected/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-kube-api-access-mt28c\") pod \"keystone-db-sync-s85wb\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:57 crc kubenswrapper[4945]: I1206 08:15:57.213819 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s85wb" Dec 06 08:15:57 crc kubenswrapper[4945]: I1206 08:15:57.649468 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-s85wb"] Dec 06 08:15:58 crc kubenswrapper[4945]: I1206 08:15:58.426099 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s85wb" event={"ID":"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3","Type":"ContainerStarted","Data":"d26b1c58c88b5a182358e9681491d6abd5493cb695228048bb702fddcbfc12ce"} Dec 06 08:16:01 crc kubenswrapper[4945]: I1206 08:16:01.212516 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 06 08:16:03 crc kubenswrapper[4945]: I1206 08:16:03.477698 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s85wb" event={"ID":"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3","Type":"ContainerStarted","Data":"1c0bc22c75aeb954f03758e5eb5a43e4189fe1d2a667b041b4ff21d91bee7eeb"} Dec 06 08:16:03 crc kubenswrapper[4945]: I1206 08:16:03.498595 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-s85wb" podStartSLOduration=2.011957724 podStartE2EDuration="7.498573685s" podCreationTimestamp="2025-12-06 08:15:56 +0000 UTC" firstStartedPulling="2025-12-06 08:15:57.657583811 +0000 UTC m=+4971.112444855" lastFinishedPulling="2025-12-06 08:16:03.144199772 +0000 UTC m=+4976.599060816" observedRunningTime="2025-12-06 08:16:03.490458988 +0000 UTC m=+4976.945320032" watchObservedRunningTime="2025-12-06 08:16:03.498573685 +0000 UTC m=+4976.953434739" Dec 06 08:16:05 crc kubenswrapper[4945]: I1206 08:16:05.497208 4945 generic.go:334] "Generic (PLEG): container finished" podID="79b89fa1-ca85-46d1-a9dd-0c15ef2052e3" containerID="1c0bc22c75aeb954f03758e5eb5a43e4189fe1d2a667b041b4ff21d91bee7eeb" exitCode=0 Dec 06 08:16:05 crc kubenswrapper[4945]: I1206 08:16:05.497322 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s85wb" event={"ID":"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3","Type":"ContainerDied","Data":"1c0bc22c75aeb954f03758e5eb5a43e4189fe1d2a667b041b4ff21d91bee7eeb"} Dec 06 08:16:06 crc kubenswrapper[4945]: I1206 08:16:06.890455 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s85wb" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.055315 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-config-data\") pod \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.055494 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-combined-ca-bundle\") pod \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.055590 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt28c\" (UniqueName: \"kubernetes.io/projected/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-kube-api-access-mt28c\") pod \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\" (UID: \"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3\") " Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.059969 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-kube-api-access-mt28c" (OuterVolumeSpecName: "kube-api-access-mt28c") pod "79b89fa1-ca85-46d1-a9dd-0c15ef2052e3" (UID: "79b89fa1-ca85-46d1-a9dd-0c15ef2052e3"). InnerVolumeSpecName "kube-api-access-mt28c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.082669 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79b89fa1-ca85-46d1-a9dd-0c15ef2052e3" (UID: "79b89fa1-ca85-46d1-a9dd-0c15ef2052e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.102117 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-config-data" (OuterVolumeSpecName: "config-data") pod "79b89fa1-ca85-46d1-a9dd-0c15ef2052e3" (UID: "79b89fa1-ca85-46d1-a9dd-0c15ef2052e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.156945 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.156981 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt28c\" (UniqueName: \"kubernetes.io/projected/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-kube-api-access-mt28c\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.156991 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.522965 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s85wb" event={"ID":"79b89fa1-ca85-46d1-a9dd-0c15ef2052e3","Type":"ContainerDied","Data":"d26b1c58c88b5a182358e9681491d6abd5493cb695228048bb702fddcbfc12ce"} Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.523014 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d26b1c58c88b5a182358e9681491d6abd5493cb695228048bb702fddcbfc12ce" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.523034 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s85wb" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.767903 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76b898c7cf-667f7"] Dec 06 08:16:07 crc kubenswrapper[4945]: E1206 08:16:07.768307 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b89fa1-ca85-46d1-a9dd-0c15ef2052e3" containerName="keystone-db-sync" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.768326 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b89fa1-ca85-46d1-a9dd-0c15ef2052e3" containerName="keystone-db-sync" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.768490 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="79b89fa1-ca85-46d1-a9dd-0c15ef2052e3" containerName="keystone-db-sync" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.769360 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.797018 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b898c7cf-667f7"] Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.868141 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phk6f\" (UniqueName: \"kubernetes.io/projected/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-kube-api-access-phk6f\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.868191 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-nb\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.868220 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-dns-svc\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.868430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-config\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.868486 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-sb\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.882668 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-tlzzn"] Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.883752 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.886581 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.886983 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.887139 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r6cjp" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.887293 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.895650 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.915070 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tlzzn"] Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.969758 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-config\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.969802 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-sb\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.969890 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phk6f\" (UniqueName: \"kubernetes.io/projected/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-kube-api-access-phk6f\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.969908 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-nb\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.969935 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-dns-svc\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.970608 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-config\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.973882 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-dns-svc\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.976809 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-nb\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.977638 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-sb\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:07 crc kubenswrapper[4945]: I1206 08:16:07.991995 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phk6f\" (UniqueName: \"kubernetes.io/projected/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-kube-api-access-phk6f\") pod \"dnsmasq-dns-76b898c7cf-667f7\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.071190 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-757dj\" (UniqueName: \"kubernetes.io/projected/595b0414-60f2-459b-b3e5-4f9c769c6610-kube-api-access-757dj\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.071238 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-credential-keys\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.071310 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-fernet-keys\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.071339 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-combined-ca-bundle\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.071442 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-config-data\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.071484 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-scripts\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.147837 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.172977 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-scripts\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.173083 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-757dj\" (UniqueName: \"kubernetes.io/projected/595b0414-60f2-459b-b3e5-4f9c769c6610-kube-api-access-757dj\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.173119 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-credential-keys\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.173162 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-fernet-keys\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.173179 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-combined-ca-bundle\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.173262 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-config-data\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.177947 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-config-data\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.180642 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-scripts\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.194685 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-combined-ca-bundle\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.196223 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-credential-keys\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.198045 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-757dj\" (UniqueName: \"kubernetes.io/projected/595b0414-60f2-459b-b3e5-4f9c769c6610-kube-api-access-757dj\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.198163 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-fernet-keys\") pod \"keystone-bootstrap-tlzzn\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.215457 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.655327 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b898c7cf-667f7"] Dec 06 08:16:08 crc kubenswrapper[4945]: W1206 08:16:08.662497 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ffa54f0_4b4a_4c9b_a8bd_f526618ec569.slice/crio-8715f134363d1a8badf921f020d7c24efaec89e91c70aa7fb32d5405f41c91d5 WatchSource:0}: Error finding container 8715f134363d1a8badf921f020d7c24efaec89e91c70aa7fb32d5405f41c91d5: Status 404 returned error can't find the container with id 8715f134363d1a8badf921f020d7c24efaec89e91c70aa7fb32d5405f41c91d5 Dec 06 08:16:08 crc kubenswrapper[4945]: I1206 08:16:08.755061 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tlzzn"] Dec 06 08:16:08 crc kubenswrapper[4945]: W1206 08:16:08.764258 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod595b0414_60f2_459b_b3e5_4f9c769c6610.slice/crio-49fb0686f8f4b5ebff6d459713b16078296c4006add93aa4fc973eb2d541c939 WatchSource:0}: Error finding container 49fb0686f8f4b5ebff6d459713b16078296c4006add93aa4fc973eb2d541c939: Status 404 returned error can't find the container with id 49fb0686f8f4b5ebff6d459713b16078296c4006add93aa4fc973eb2d541c939 Dec 06 08:16:09 crc kubenswrapper[4945]: I1206 08:16:09.541982 4945 generic.go:334] "Generic (PLEG): container finished" podID="6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" containerID="5493a13249c18257ae4481e5af502376ea5cf825e31781b3fdeed683101c4acf" exitCode=0 Dec 06 08:16:09 crc kubenswrapper[4945]: I1206 08:16:09.542072 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" event={"ID":"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569","Type":"ContainerDied","Data":"5493a13249c18257ae4481e5af502376ea5cf825e31781b3fdeed683101c4acf"} Dec 06 08:16:09 crc kubenswrapper[4945]: I1206 08:16:09.543576 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" event={"ID":"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569","Type":"ContainerStarted","Data":"8715f134363d1a8badf921f020d7c24efaec89e91c70aa7fb32d5405f41c91d5"} Dec 06 08:16:09 crc kubenswrapper[4945]: I1206 08:16:09.547434 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tlzzn" event={"ID":"595b0414-60f2-459b-b3e5-4f9c769c6610","Type":"ContainerStarted","Data":"b91f7825d10936ccc7c298ec92cce4b1cb2c001b18565b7189aca2484035f00f"} Dec 06 08:16:09 crc kubenswrapper[4945]: I1206 08:16:09.547492 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tlzzn" event={"ID":"595b0414-60f2-459b-b3e5-4f9c769c6610","Type":"ContainerStarted","Data":"49fb0686f8f4b5ebff6d459713b16078296c4006add93aa4fc973eb2d541c939"} Dec 06 08:16:09 crc kubenswrapper[4945]: I1206 08:16:09.596182 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-tlzzn" podStartSLOduration=2.596163399 podStartE2EDuration="2.596163399s" podCreationTimestamp="2025-12-06 08:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:16:09.590165849 +0000 UTC m=+4983.045026903" watchObservedRunningTime="2025-12-06 08:16:09.596163399 +0000 UTC m=+4983.051024443" Dec 06 08:16:10 crc kubenswrapper[4945]: I1206 08:16:10.557144 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" event={"ID":"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569","Type":"ContainerStarted","Data":"6f2fe87aec157c5fa0c110ce5891adb88be8cea215dbdf8fea66d28874545840"} Dec 06 08:16:10 crc kubenswrapper[4945]: I1206 08:16:10.584873 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" podStartSLOduration=3.584845129 podStartE2EDuration="3.584845129s" podCreationTimestamp="2025-12-06 08:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:16:10.577572455 +0000 UTC m=+4984.032433499" watchObservedRunningTime="2025-12-06 08:16:10.584845129 +0000 UTC m=+4984.039706193" Dec 06 08:16:11 crc kubenswrapper[4945]: I1206 08:16:11.563492 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:12 crc kubenswrapper[4945]: I1206 08:16:12.571762 4945 generic.go:334] "Generic (PLEG): container finished" podID="595b0414-60f2-459b-b3e5-4f9c769c6610" containerID="b91f7825d10936ccc7c298ec92cce4b1cb2c001b18565b7189aca2484035f00f" exitCode=0 Dec 06 08:16:12 crc kubenswrapper[4945]: I1206 08:16:12.571848 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tlzzn" event={"ID":"595b0414-60f2-459b-b3e5-4f9c769c6610","Type":"ContainerDied","Data":"b91f7825d10936ccc7c298ec92cce4b1cb2c001b18565b7189aca2484035f00f"} Dec 06 08:16:13 crc kubenswrapper[4945]: I1206 08:16:13.948911 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.091192 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-scripts\") pod \"595b0414-60f2-459b-b3e5-4f9c769c6610\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.091373 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-combined-ca-bundle\") pod \"595b0414-60f2-459b-b3e5-4f9c769c6610\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.091428 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-757dj\" (UniqueName: \"kubernetes.io/projected/595b0414-60f2-459b-b3e5-4f9c769c6610-kube-api-access-757dj\") pod \"595b0414-60f2-459b-b3e5-4f9c769c6610\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.091492 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-credential-keys\") pod \"595b0414-60f2-459b-b3e5-4f9c769c6610\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.091553 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-config-data\") pod \"595b0414-60f2-459b-b3e5-4f9c769c6610\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.091587 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-fernet-keys\") pod \"595b0414-60f2-459b-b3e5-4f9c769c6610\" (UID: \"595b0414-60f2-459b-b3e5-4f9c769c6610\") " Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.096790 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "595b0414-60f2-459b-b3e5-4f9c769c6610" (UID: "595b0414-60f2-459b-b3e5-4f9c769c6610"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.097126 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/595b0414-60f2-459b-b3e5-4f9c769c6610-kube-api-access-757dj" (OuterVolumeSpecName: "kube-api-access-757dj") pod "595b0414-60f2-459b-b3e5-4f9c769c6610" (UID: "595b0414-60f2-459b-b3e5-4f9c769c6610"). InnerVolumeSpecName "kube-api-access-757dj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.097444 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "595b0414-60f2-459b-b3e5-4f9c769c6610" (UID: "595b0414-60f2-459b-b3e5-4f9c769c6610"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.099408 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-scripts" (OuterVolumeSpecName: "scripts") pod "595b0414-60f2-459b-b3e5-4f9c769c6610" (UID: "595b0414-60f2-459b-b3e5-4f9c769c6610"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.115014 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "595b0414-60f2-459b-b3e5-4f9c769c6610" (UID: "595b0414-60f2-459b-b3e5-4f9c769c6610"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.125585 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-config-data" (OuterVolumeSpecName: "config-data") pod "595b0414-60f2-459b-b3e5-4f9c769c6610" (UID: "595b0414-60f2-459b-b3e5-4f9c769c6610"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.194436 4945 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.194491 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.194511 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.194527 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.194545 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595b0414-60f2-459b-b3e5-4f9c769c6610-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.194566 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-757dj\" (UniqueName: \"kubernetes.io/projected/595b0414-60f2-459b-b3e5-4f9c769c6610-kube-api-access-757dj\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.589731 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tlzzn" event={"ID":"595b0414-60f2-459b-b3e5-4f9c769c6610","Type":"ContainerDied","Data":"49fb0686f8f4b5ebff6d459713b16078296c4006add93aa4fc973eb2d541c939"} Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.590031 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49fb0686f8f4b5ebff6d459713b16078296c4006add93aa4fc973eb2d541c939" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.589786 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tlzzn" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.663552 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-tlzzn"] Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.671874 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-tlzzn"] Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.753046 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-t6tq5"] Dec 06 08:16:14 crc kubenswrapper[4945]: E1206 08:16:14.753423 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="595b0414-60f2-459b-b3e5-4f9c769c6610" containerName="keystone-bootstrap" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.753445 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="595b0414-60f2-459b-b3e5-4f9c769c6610" containerName="keystone-bootstrap" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.753654 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="595b0414-60f2-459b-b3e5-4f9c769c6610" containerName="keystone-bootstrap" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.754371 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.757296 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.757424 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r6cjp" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.757501 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.757781 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.757950 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.767066 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t6tq5"] Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.908626 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-config-data\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.908678 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-combined-ca-bundle\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.908707 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-scripts\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.908879 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm5zb\" (UniqueName: \"kubernetes.io/projected/c16c276f-ad77-45ab-b719-368517cf919d-kube-api-access-jm5zb\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.908919 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-fernet-keys\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.908967 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-credential-keys\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:14 crc kubenswrapper[4945]: I1206 08:16:14.964652 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="595b0414-60f2-459b-b3e5-4f9c769c6610" path="/var/lib/kubelet/pods/595b0414-60f2-459b-b3e5-4f9c769c6610/volumes" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.010909 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm5zb\" (UniqueName: \"kubernetes.io/projected/c16c276f-ad77-45ab-b719-368517cf919d-kube-api-access-jm5zb\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.010957 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-fernet-keys\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.010994 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-credential-keys\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.011053 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-config-data\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.011069 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-combined-ca-bundle\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.011084 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-scripts\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.015317 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-config-data\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.016055 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-credential-keys\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.016459 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-fernet-keys\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.019833 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-combined-ca-bundle\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.020858 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-scripts\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.028965 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm5zb\" (UniqueName: \"kubernetes.io/projected/c16c276f-ad77-45ab-b719-368517cf919d-kube-api-access-jm5zb\") pod \"keystone-bootstrap-t6tq5\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.072534 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:15 crc kubenswrapper[4945]: I1206 08:16:15.646392 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t6tq5"] Dec 06 08:16:16 crc kubenswrapper[4945]: I1206 08:16:16.609911 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t6tq5" event={"ID":"c16c276f-ad77-45ab-b719-368517cf919d","Type":"ContainerStarted","Data":"e5bd8b31440c0e0732c0f41fe457000b4c885d5cd8e02288919d80fbb4a80801"} Dec 06 08:16:16 crc kubenswrapper[4945]: I1206 08:16:16.610323 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t6tq5" event={"ID":"c16c276f-ad77-45ab-b719-368517cf919d","Type":"ContainerStarted","Data":"1da6829fe6d1ef873b50feec26a70eff69d071acc1527cf082f13a54e40d9a7f"} Dec 06 08:16:16 crc kubenswrapper[4945]: I1206 08:16:16.631073 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-t6tq5" podStartSLOduration=2.631056669 podStartE2EDuration="2.631056669s" podCreationTimestamp="2025-12-06 08:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:16:16.62696608 +0000 UTC m=+4990.081827144" watchObservedRunningTime="2025-12-06 08:16:16.631056669 +0000 UTC m=+4990.085917713" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.149882 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.225153 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65bf94dc4c-szkv6"] Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.231418 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" podUID="526d2c6c-3a06-44e8-8def-4ef3fff34ac9" containerName="dnsmasq-dns" containerID="cri-o://c9de46e2658c6071dc8376f6b0acc2d5183c3f17f8b3bd5b7a5efc36d51e8aba" gracePeriod=10 Dec 06 08:16:18 crc kubenswrapper[4945]: E1206 08:16:18.464249 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod526d2c6c_3a06_44e8_8def_4ef3fff34ac9.slice/crio-conmon-c9de46e2658c6071dc8376f6b0acc2d5183c3f17f8b3bd5b7a5efc36d51e8aba.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod526d2c6c_3a06_44e8_8def_4ef3fff34ac9.slice/crio-c9de46e2658c6071dc8376f6b0acc2d5183c3f17f8b3bd5b7a5efc36d51e8aba.scope\": RecentStats: unable to find data in memory cache]" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.655664 4945 generic.go:334] "Generic (PLEG): container finished" podID="526d2c6c-3a06-44e8-8def-4ef3fff34ac9" containerID="c9de46e2658c6071dc8376f6b0acc2d5183c3f17f8b3bd5b7a5efc36d51e8aba" exitCode=0 Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.656067 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" event={"ID":"526d2c6c-3a06-44e8-8def-4ef3fff34ac9","Type":"ContainerDied","Data":"c9de46e2658c6071dc8376f6b0acc2d5183c3f17f8b3bd5b7a5efc36d51e8aba"} Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.656102 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" event={"ID":"526d2c6c-3a06-44e8-8def-4ef3fff34ac9","Type":"ContainerDied","Data":"f9ced3d73ca847d5588a939e670beb23000be3761e294756649c94db6d23c772"} Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.656115 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9ced3d73ca847d5588a939e670beb23000be3761e294756649c94db6d23c772" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.670516 4945 generic.go:334] "Generic (PLEG): container finished" podID="c16c276f-ad77-45ab-b719-368517cf919d" containerID="e5bd8b31440c0e0732c0f41fe457000b4c885d5cd8e02288919d80fbb4a80801" exitCode=0 Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.670568 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t6tq5" event={"ID":"c16c276f-ad77-45ab-b719-368517cf919d","Type":"ContainerDied","Data":"e5bd8b31440c0e0732c0f41fe457000b4c885d5cd8e02288919d80fbb4a80801"} Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.704219 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.885460 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-dns-svc\") pod \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.885642 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-sb\") pod \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.885712 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-config\") pod \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.885748 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fplc\" (UniqueName: \"kubernetes.io/projected/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-kube-api-access-4fplc\") pod \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.885822 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-nb\") pod \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\" (UID: \"526d2c6c-3a06-44e8-8def-4ef3fff34ac9\") " Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.896561 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-kube-api-access-4fplc" (OuterVolumeSpecName: "kube-api-access-4fplc") pod "526d2c6c-3a06-44e8-8def-4ef3fff34ac9" (UID: "526d2c6c-3a06-44e8-8def-4ef3fff34ac9"). InnerVolumeSpecName "kube-api-access-4fplc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.929834 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "526d2c6c-3a06-44e8-8def-4ef3fff34ac9" (UID: "526d2c6c-3a06-44e8-8def-4ef3fff34ac9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.929921 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "526d2c6c-3a06-44e8-8def-4ef3fff34ac9" (UID: "526d2c6c-3a06-44e8-8def-4ef3fff34ac9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.931189 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "526d2c6c-3a06-44e8-8def-4ef3fff34ac9" (UID: "526d2c6c-3a06-44e8-8def-4ef3fff34ac9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.944727 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-config" (OuterVolumeSpecName: "config") pod "526d2c6c-3a06-44e8-8def-4ef3fff34ac9" (UID: "526d2c6c-3a06-44e8-8def-4ef3fff34ac9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.988108 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.988137 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.988147 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fplc\" (UniqueName: \"kubernetes.io/projected/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-kube-api-access-4fplc\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.988156 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:18 crc kubenswrapper[4945]: I1206 08:16:18.988165 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/526d2c6c-3a06-44e8-8def-4ef3fff34ac9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:19 crc kubenswrapper[4945]: I1206 08:16:19.679946 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bf94dc4c-szkv6" Dec 06 08:16:19 crc kubenswrapper[4945]: I1206 08:16:19.709121 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65bf94dc4c-szkv6"] Dec 06 08:16:19 crc kubenswrapper[4945]: I1206 08:16:19.714623 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65bf94dc4c-szkv6"] Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.005796 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.106527 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm5zb\" (UniqueName: \"kubernetes.io/projected/c16c276f-ad77-45ab-b719-368517cf919d-kube-api-access-jm5zb\") pod \"c16c276f-ad77-45ab-b719-368517cf919d\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.106625 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-combined-ca-bundle\") pod \"c16c276f-ad77-45ab-b719-368517cf919d\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.106725 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-fernet-keys\") pod \"c16c276f-ad77-45ab-b719-368517cf919d\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.106747 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-config-data\") pod \"c16c276f-ad77-45ab-b719-368517cf919d\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.106776 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-scripts\") pod \"c16c276f-ad77-45ab-b719-368517cf919d\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.106807 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-credential-keys\") pod \"c16c276f-ad77-45ab-b719-368517cf919d\" (UID: \"c16c276f-ad77-45ab-b719-368517cf919d\") " Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.111436 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c16c276f-ad77-45ab-b719-368517cf919d" (UID: "c16c276f-ad77-45ab-b719-368517cf919d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.112009 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c16c276f-ad77-45ab-b719-368517cf919d-kube-api-access-jm5zb" (OuterVolumeSpecName: "kube-api-access-jm5zb") pod "c16c276f-ad77-45ab-b719-368517cf919d" (UID: "c16c276f-ad77-45ab-b719-368517cf919d"). InnerVolumeSpecName "kube-api-access-jm5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.114573 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-scripts" (OuterVolumeSpecName: "scripts") pod "c16c276f-ad77-45ab-b719-368517cf919d" (UID: "c16c276f-ad77-45ab-b719-368517cf919d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.114851 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c16c276f-ad77-45ab-b719-368517cf919d" (UID: "c16c276f-ad77-45ab-b719-368517cf919d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.139734 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-config-data" (OuterVolumeSpecName: "config-data") pod "c16c276f-ad77-45ab-b719-368517cf919d" (UID: "c16c276f-ad77-45ab-b719-368517cf919d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.147196 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c16c276f-ad77-45ab-b719-368517cf919d" (UID: "c16c276f-ad77-45ab-b719-368517cf919d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.208713 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.208757 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.208767 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.208776 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.208784 4945 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c16c276f-ad77-45ab-b719-368517cf919d-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.208792 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm5zb\" (UniqueName: \"kubernetes.io/projected/c16c276f-ad77-45ab-b719-368517cf919d-kube-api-access-jm5zb\") on node \"crc\" DevicePath \"\"" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.688973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t6tq5" event={"ID":"c16c276f-ad77-45ab-b719-368517cf919d","Type":"ContainerDied","Data":"1da6829fe6d1ef873b50feec26a70eff69d071acc1527cf082f13a54e40d9a7f"} Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.689037 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1da6829fe6d1ef873b50feec26a70eff69d071acc1527cf082f13a54e40d9a7f" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.689044 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t6tq5" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.796494 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c87cbcc4c-8jnjw"] Dec 06 08:16:20 crc kubenswrapper[4945]: E1206 08:16:20.796880 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16c276f-ad77-45ab-b719-368517cf919d" containerName="keystone-bootstrap" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.796907 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16c276f-ad77-45ab-b719-368517cf919d" containerName="keystone-bootstrap" Dec 06 08:16:20 crc kubenswrapper[4945]: E1206 08:16:20.796928 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="526d2c6c-3a06-44e8-8def-4ef3fff34ac9" containerName="init" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.796936 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="526d2c6c-3a06-44e8-8def-4ef3fff34ac9" containerName="init" Dec 06 08:16:20 crc kubenswrapper[4945]: E1206 08:16:20.796973 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="526d2c6c-3a06-44e8-8def-4ef3fff34ac9" containerName="dnsmasq-dns" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.796981 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="526d2c6c-3a06-44e8-8def-4ef3fff34ac9" containerName="dnsmasq-dns" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.797185 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16c276f-ad77-45ab-b719-368517cf919d" containerName="keystone-bootstrap" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.797210 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="526d2c6c-3a06-44e8-8def-4ef3fff34ac9" containerName="dnsmasq-dns" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.798138 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.799889 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r6cjp" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.801603 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.801793 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.802847 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.805002 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c87cbcc4c-8jnjw"] Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.919567 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-config-data\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.919618 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-fernet-keys\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.919650 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-credential-keys\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.919787 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-scripts\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.919833 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lz2b\" (UniqueName: \"kubernetes.io/projected/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-kube-api-access-2lz2b\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.919869 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-combined-ca-bundle\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:20 crc kubenswrapper[4945]: I1206 08:16:20.968784 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="526d2c6c-3a06-44e8-8def-4ef3fff34ac9" path="/var/lib/kubelet/pods/526d2c6c-3a06-44e8-8def-4ef3fff34ac9/volumes" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.021605 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-combined-ca-bundle\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.021714 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-config-data\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.021737 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-fernet-keys\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.021756 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-credential-keys\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.021815 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-scripts\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.021844 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lz2b\" (UniqueName: \"kubernetes.io/projected/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-kube-api-access-2lz2b\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.025661 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-fernet-keys\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.025887 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-config-data\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.026549 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-combined-ca-bundle\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.026558 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-scripts\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.026909 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-credential-keys\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.035889 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lz2b\" (UniqueName: \"kubernetes.io/projected/d5e3e5df-4483-49ee-9d22-cf9d3a8757c1-kube-api-access-2lz2b\") pod \"keystone-6c87cbcc4c-8jnjw\" (UID: \"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1\") " pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.119055 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.529556 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c87cbcc4c-8jnjw"] Dec 06 08:16:21 crc kubenswrapper[4945]: I1206 08:16:21.700310 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c87cbcc4c-8jnjw" event={"ID":"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1","Type":"ContainerStarted","Data":"5eca53491d815007caf022920c400f8b150623933feacd2af5faf7dbaa0cf223"} Dec 06 08:16:22 crc kubenswrapper[4945]: I1206 08:16:22.708695 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c87cbcc4c-8jnjw" event={"ID":"d5e3e5df-4483-49ee-9d22-cf9d3a8757c1","Type":"ContainerStarted","Data":"59059c35771fec613aa270d668ed40d156e360002fe8f845421b3273b2d8ec9f"} Dec 06 08:16:22 crc kubenswrapper[4945]: I1206 08:16:22.708893 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:22 crc kubenswrapper[4945]: I1206 08:16:22.744113 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c87cbcc4c-8jnjw" podStartSLOduration=2.744092506 podStartE2EDuration="2.744092506s" podCreationTimestamp="2025-12-06 08:16:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:16:22.723720561 +0000 UTC m=+4996.178581605" watchObservedRunningTime="2025-12-06 08:16:22.744092506 +0000 UTC m=+4996.198953550" Dec 06 08:16:38 crc kubenswrapper[4945]: I1206 08:16:38.795841 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:16:38 crc kubenswrapper[4945]: I1206 08:16:38.796659 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:16:52 crc kubenswrapper[4945]: I1206 08:16:52.586646 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c87cbcc4c-8jnjw" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.362695 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.365822 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.370032 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.370649 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-bhrl4" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.371093 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.372231 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.469957 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config\") pod \"openstackclient\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " pod="openstack/openstackclient" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.470157 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bclr7\" (UniqueName: \"kubernetes.io/projected/039bb79d-fba0-413e-844f-e21e8ec625d6-kube-api-access-bclr7\") pod \"openstackclient\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " pod="openstack/openstackclient" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.470185 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config-secret\") pod \"openstackclient\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " pod="openstack/openstackclient" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.572084 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bclr7\" (UniqueName: \"kubernetes.io/projected/039bb79d-fba0-413e-844f-e21e8ec625d6-kube-api-access-bclr7\") pod \"openstackclient\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " pod="openstack/openstackclient" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.572156 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config-secret\") pod \"openstackclient\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " pod="openstack/openstackclient" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.572296 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config\") pod \"openstackclient\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " pod="openstack/openstackclient" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.573388 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config\") pod \"openstackclient\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " pod="openstack/openstackclient" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.592181 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config-secret\") pod \"openstackclient\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " pod="openstack/openstackclient" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.593916 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bclr7\" (UniqueName: \"kubernetes.io/projected/039bb79d-fba0-413e-844f-e21e8ec625d6-kube-api-access-bclr7\") pod \"openstackclient\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " pod="openstack/openstackclient" Dec 06 08:16:56 crc kubenswrapper[4945]: I1206 08:16:56.705733 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 08:16:57 crc kubenswrapper[4945]: I1206 08:16:57.182924 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 08:16:58 crc kubenswrapper[4945]: I1206 08:16:58.027851 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"039bb79d-fba0-413e-844f-e21e8ec625d6","Type":"ContainerStarted","Data":"2030ff8048d055756abbab2ff277ba10ffca61fbd98255c9a47bd1b97aa67ec2"} Dec 06 08:17:08 crc kubenswrapper[4945]: I1206 08:17:08.795732 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:17:08 crc kubenswrapper[4945]: I1206 08:17:08.796245 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:17:09 crc kubenswrapper[4945]: I1206 08:17:09.110272 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"039bb79d-fba0-413e-844f-e21e8ec625d6","Type":"ContainerStarted","Data":"8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61"} Dec 06 08:17:09 crc kubenswrapper[4945]: I1206 08:17:09.133234 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.142723461 podStartE2EDuration="13.133217023s" podCreationTimestamp="2025-12-06 08:16:56 +0000 UTC" firstStartedPulling="2025-12-06 08:16:57.192825657 +0000 UTC m=+5030.647686701" lastFinishedPulling="2025-12-06 08:17:08.183319219 +0000 UTC m=+5041.638180263" observedRunningTime="2025-12-06 08:17:09.131472886 +0000 UTC m=+5042.586333970" watchObservedRunningTime="2025-12-06 08:17:09.133217023 +0000 UTC m=+5042.588078067" Dec 06 08:17:38 crc kubenswrapper[4945]: I1206 08:17:38.795547 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:17:38 crc kubenswrapper[4945]: I1206 08:17:38.796355 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:17:38 crc kubenswrapper[4945]: I1206 08:17:38.796463 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 08:17:38 crc kubenswrapper[4945]: I1206 08:17:38.797042 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e73a93782f4167b2d20799f17eea711a9f4910590603bc3ce04323932093a55"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 08:17:38 crc kubenswrapper[4945]: I1206 08:17:38.797120 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://8e73a93782f4167b2d20799f17eea711a9f4910590603bc3ce04323932093a55" gracePeriod=600 Dec 06 08:17:40 crc kubenswrapper[4945]: I1206 08:17:40.380939 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="8e73a93782f4167b2d20799f17eea711a9f4910590603bc3ce04323932093a55" exitCode=0 Dec 06 08:17:40 crc kubenswrapper[4945]: I1206 08:17:40.381045 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"8e73a93782f4167b2d20799f17eea711a9f4910590603bc3ce04323932093a55"} Dec 06 08:17:40 crc kubenswrapper[4945]: I1206 08:17:40.381233 4945 scope.go:117] "RemoveContainer" containerID="5e67cd00fee46711c697ca0bb434d71c6a797b86f3f4dcf115a1a1a1c05204d4" Dec 06 08:17:41 crc kubenswrapper[4945]: I1206 08:17:41.391527 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39"} Dec 06 08:18:24 crc kubenswrapper[4945]: I1206 08:18:24.440080 4945 scope.go:117] "RemoveContainer" containerID="4c5c07e597bb3b32880df7fba31c8a6eb5ccf842e12c452883deeefed88b9738" Dec 06 08:18:24 crc kubenswrapper[4945]: I1206 08:18:24.460675 4945 scope.go:117] "RemoveContainer" containerID="170bcfacc1df27ed3fc03ecce2e09eab392d39b3f533ff6421b6412dc5a0cd43" Dec 06 08:18:24 crc kubenswrapper[4945]: I1206 08:18:24.498917 4945 scope.go:117] "RemoveContainer" containerID="cf4d41f3552b6cbef44bd18840a6f336a53ffc376a9eab51d8954ddd4131d890" Dec 06 08:18:24 crc kubenswrapper[4945]: I1206 08:18:24.536374 4945 scope.go:117] "RemoveContainer" containerID="50ca7d9afaaa496cc1db00cc3dcac5da2d8756f006df2b1b548b5fc2ebbc74bc" Dec 06 08:18:24 crc kubenswrapper[4945]: I1206 08:18:24.561805 4945 scope.go:117] "RemoveContainer" containerID="3f4b7edabf35d1e03e6a921b2107b26b4eda2700b08cbea8040a43d2d030d1e0" Dec 06 08:18:24 crc kubenswrapper[4945]: I1206 08:18:24.589755 4945 scope.go:117] "RemoveContainer" containerID="14a5568c03a6f199e2529308325d625bd697ebbc72a1d77ecdcffd846126a0bd" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.475152 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-v879n"] Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.477269 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v879n" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.492485 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-39b5-account-create-update-6l9r9"] Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.493787 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-39b5-account-create-update-6l9r9" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.496347 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.501671 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-v879n"] Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.508695 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-39b5-account-create-update-6l9r9"] Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.584929 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-operator-scripts\") pod \"barbican-39b5-account-create-update-6l9r9\" (UID: \"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9\") " pod="openstack/barbican-39b5-account-create-update-6l9r9" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.585104 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzdr8\" (UniqueName: \"kubernetes.io/projected/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-kube-api-access-nzdr8\") pod \"barbican-39b5-account-create-update-6l9r9\" (UID: \"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9\") " pod="openstack/barbican-39b5-account-create-update-6l9r9" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.585396 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de99182f-85eb-43ec-a5a5-c5ec7779b399-operator-scripts\") pod \"barbican-db-create-v879n\" (UID: \"de99182f-85eb-43ec-a5a5-c5ec7779b399\") " pod="openstack/barbican-db-create-v879n" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.585704 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8bhl\" (UniqueName: \"kubernetes.io/projected/de99182f-85eb-43ec-a5a5-c5ec7779b399-kube-api-access-p8bhl\") pod \"barbican-db-create-v879n\" (UID: \"de99182f-85eb-43ec-a5a5-c5ec7779b399\") " pod="openstack/barbican-db-create-v879n" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.686877 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de99182f-85eb-43ec-a5a5-c5ec7779b399-operator-scripts\") pod \"barbican-db-create-v879n\" (UID: \"de99182f-85eb-43ec-a5a5-c5ec7779b399\") " pod="openstack/barbican-db-create-v879n" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.686973 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8bhl\" (UniqueName: \"kubernetes.io/projected/de99182f-85eb-43ec-a5a5-c5ec7779b399-kube-api-access-p8bhl\") pod \"barbican-db-create-v879n\" (UID: \"de99182f-85eb-43ec-a5a5-c5ec7779b399\") " pod="openstack/barbican-db-create-v879n" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.686997 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-operator-scripts\") pod \"barbican-39b5-account-create-update-6l9r9\" (UID: \"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9\") " pod="openstack/barbican-39b5-account-create-update-6l9r9" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.687021 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzdr8\" (UniqueName: \"kubernetes.io/projected/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-kube-api-access-nzdr8\") pod \"barbican-39b5-account-create-update-6l9r9\" (UID: \"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9\") " pod="openstack/barbican-39b5-account-create-update-6l9r9" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.687938 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de99182f-85eb-43ec-a5a5-c5ec7779b399-operator-scripts\") pod \"barbican-db-create-v879n\" (UID: \"de99182f-85eb-43ec-a5a5-c5ec7779b399\") " pod="openstack/barbican-db-create-v879n" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.688799 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-operator-scripts\") pod \"barbican-39b5-account-create-update-6l9r9\" (UID: \"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9\") " pod="openstack/barbican-39b5-account-create-update-6l9r9" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.705305 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzdr8\" (UniqueName: \"kubernetes.io/projected/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-kube-api-access-nzdr8\") pod \"barbican-39b5-account-create-update-6l9r9\" (UID: \"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9\") " pod="openstack/barbican-39b5-account-create-update-6l9r9" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.705791 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8bhl\" (UniqueName: \"kubernetes.io/projected/de99182f-85eb-43ec-a5a5-c5ec7779b399-kube-api-access-p8bhl\") pod \"barbican-db-create-v879n\" (UID: \"de99182f-85eb-43ec-a5a5-c5ec7779b399\") " pod="openstack/barbican-db-create-v879n" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.850701 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v879n" Dec 06 08:18:32 crc kubenswrapper[4945]: I1206 08:18:32.860965 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-39b5-account-create-update-6l9r9" Dec 06 08:18:33 crc kubenswrapper[4945]: I1206 08:18:33.293247 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-v879n"] Dec 06 08:18:33 crc kubenswrapper[4945]: I1206 08:18:33.352687 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-39b5-account-create-update-6l9r9"] Dec 06 08:18:33 crc kubenswrapper[4945]: W1206 08:18:33.357432 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c2ce74f_0f26_4db5_b6b4_10ebb3ad80f9.slice/crio-8d1babffcc58ab41482d37d4028d69d361aed18453fb58566586a07d8d1b55b0 WatchSource:0}: Error finding container 8d1babffcc58ab41482d37d4028d69d361aed18453fb58566586a07d8d1b55b0: Status 404 returned error can't find the container with id 8d1babffcc58ab41482d37d4028d69d361aed18453fb58566586a07d8d1b55b0 Dec 06 08:18:33 crc kubenswrapper[4945]: I1206 08:18:33.832218 4945 generic.go:334] "Generic (PLEG): container finished" podID="de99182f-85eb-43ec-a5a5-c5ec7779b399" containerID="4c015654cc5d115f268eeeb4b314d7df0e9c0f967efd273d9f4d15c824cc6864" exitCode=0 Dec 06 08:18:33 crc kubenswrapper[4945]: I1206 08:18:33.832412 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v879n" event={"ID":"de99182f-85eb-43ec-a5a5-c5ec7779b399","Type":"ContainerDied","Data":"4c015654cc5d115f268eeeb4b314d7df0e9c0f967efd273d9f4d15c824cc6864"} Dec 06 08:18:33 crc kubenswrapper[4945]: I1206 08:18:33.832447 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v879n" event={"ID":"de99182f-85eb-43ec-a5a5-c5ec7779b399","Type":"ContainerStarted","Data":"27bd1c48f62a9ee8ff2785f31c9aa37a1b34258ef2d146377c46a5cec9ab1de2"} Dec 06 08:18:33 crc kubenswrapper[4945]: I1206 08:18:33.834967 4945 generic.go:334] "Generic (PLEG): container finished" podID="9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9" containerID="8bc9b2b9b06e8a3035e1111ea4fed8b6559b28e9f717a825bad06b2c07582dda" exitCode=0 Dec 06 08:18:33 crc kubenswrapper[4945]: I1206 08:18:33.835012 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-39b5-account-create-update-6l9r9" event={"ID":"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9","Type":"ContainerDied","Data":"8bc9b2b9b06e8a3035e1111ea4fed8b6559b28e9f717a825bad06b2c07582dda"} Dec 06 08:18:33 crc kubenswrapper[4945]: I1206 08:18:33.835043 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-39b5-account-create-update-6l9r9" event={"ID":"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9","Type":"ContainerStarted","Data":"8d1babffcc58ab41482d37d4028d69d361aed18453fb58566586a07d8d1b55b0"} Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.252215 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-39b5-account-create-update-6l9r9" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.259526 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v879n" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.336566 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de99182f-85eb-43ec-a5a5-c5ec7779b399-operator-scripts\") pod \"de99182f-85eb-43ec-a5a5-c5ec7779b399\" (UID: \"de99182f-85eb-43ec-a5a5-c5ec7779b399\") " Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.336737 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8bhl\" (UniqueName: \"kubernetes.io/projected/de99182f-85eb-43ec-a5a5-c5ec7779b399-kube-api-access-p8bhl\") pod \"de99182f-85eb-43ec-a5a5-c5ec7779b399\" (UID: \"de99182f-85eb-43ec-a5a5-c5ec7779b399\") " Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.336782 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-operator-scripts\") pod \"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9\" (UID: \"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9\") " Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.336799 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzdr8\" (UniqueName: \"kubernetes.io/projected/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-kube-api-access-nzdr8\") pod \"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9\" (UID: \"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9\") " Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.337668 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9" (UID: "9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.338477 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de99182f-85eb-43ec-a5a5-c5ec7779b399-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "de99182f-85eb-43ec-a5a5-c5ec7779b399" (UID: "de99182f-85eb-43ec-a5a5-c5ec7779b399"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.344015 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de99182f-85eb-43ec-a5a5-c5ec7779b399-kube-api-access-p8bhl" (OuterVolumeSpecName: "kube-api-access-p8bhl") pod "de99182f-85eb-43ec-a5a5-c5ec7779b399" (UID: "de99182f-85eb-43ec-a5a5-c5ec7779b399"). InnerVolumeSpecName "kube-api-access-p8bhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.344796 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-kube-api-access-nzdr8" (OuterVolumeSpecName: "kube-api-access-nzdr8") pod "9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9" (UID: "9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9"). InnerVolumeSpecName "kube-api-access-nzdr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.351193 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tzkxz"] Dec 06 08:18:35 crc kubenswrapper[4945]: E1206 08:18:35.351591 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9" containerName="mariadb-account-create-update" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.352021 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9" containerName="mariadb-account-create-update" Dec 06 08:18:35 crc kubenswrapper[4945]: E1206 08:18:35.352122 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de99182f-85eb-43ec-a5a5-c5ec7779b399" containerName="mariadb-database-create" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.352177 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="de99182f-85eb-43ec-a5a5-c5ec7779b399" containerName="mariadb-database-create" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.352406 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="de99182f-85eb-43ec-a5a5-c5ec7779b399" containerName="mariadb-database-create" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.352483 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9" containerName="mariadb-account-create-update" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.355727 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.367544 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tzkxz"] Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.438581 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlr8q\" (UniqueName: \"kubernetes.io/projected/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-kube-api-access-wlr8q\") pod \"redhat-marketplace-tzkxz\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.438751 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-utilities\") pod \"redhat-marketplace-tzkxz\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.439049 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-catalog-content\") pod \"redhat-marketplace-tzkxz\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.439188 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8bhl\" (UniqueName: \"kubernetes.io/projected/de99182f-85eb-43ec-a5a5-c5ec7779b399-kube-api-access-p8bhl\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.439207 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.439218 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzdr8\" (UniqueName: \"kubernetes.io/projected/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9-kube-api-access-nzdr8\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.439227 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de99182f-85eb-43ec-a5a5-c5ec7779b399-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.540584 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-catalog-content\") pod \"redhat-marketplace-tzkxz\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.540794 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlr8q\" (UniqueName: \"kubernetes.io/projected/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-kube-api-access-wlr8q\") pod \"redhat-marketplace-tzkxz\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.540833 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-utilities\") pod \"redhat-marketplace-tzkxz\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.541232 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-catalog-content\") pod \"redhat-marketplace-tzkxz\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.541325 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-utilities\") pod \"redhat-marketplace-tzkxz\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.563430 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlr8q\" (UniqueName: \"kubernetes.io/projected/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-kube-api-access-wlr8q\") pod \"redhat-marketplace-tzkxz\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.741106 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.868264 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v879n" event={"ID":"de99182f-85eb-43ec-a5a5-c5ec7779b399","Type":"ContainerDied","Data":"27bd1c48f62a9ee8ff2785f31c9aa37a1b34258ef2d146377c46a5cec9ab1de2"} Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.868613 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27bd1c48f62a9ee8ff2785f31c9aa37a1b34258ef2d146377c46a5cec9ab1de2" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.868705 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v879n" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.877561 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-39b5-account-create-update-6l9r9" event={"ID":"9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9","Type":"ContainerDied","Data":"8d1babffcc58ab41482d37d4028d69d361aed18453fb58566586a07d8d1b55b0"} Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.877605 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d1babffcc58ab41482d37d4028d69d361aed18453fb58566586a07d8d1b55b0" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.877664 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-39b5-account-create-update-6l9r9" Dec 06 08:18:35 crc kubenswrapper[4945]: I1206 08:18:35.993484 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tzkxz"] Dec 06 08:18:35 crc kubenswrapper[4945]: W1206 08:18:35.996081 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15f29b5a_e0ef_4265_a2dc_8f159e5a8a7e.slice/crio-a51a8c5fd8acb47bcf3b387235818493f0330f137ac84b22eea5073864a6dada WatchSource:0}: Error finding container a51a8c5fd8acb47bcf3b387235818493f0330f137ac84b22eea5073864a6dada: Status 404 returned error can't find the container with id a51a8c5fd8acb47bcf3b387235818493f0330f137ac84b22eea5073864a6dada Dec 06 08:18:36 crc kubenswrapper[4945]: I1206 08:18:36.888765 4945 generic.go:334] "Generic (PLEG): container finished" podID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerID="c512cd3d350fdf15bad5dd7cb545cf61eb7b42be9a0a1fe9ea5945f04b2e23ff" exitCode=0 Dec 06 08:18:36 crc kubenswrapper[4945]: I1206 08:18:36.889146 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzkxz" event={"ID":"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e","Type":"ContainerDied","Data":"c512cd3d350fdf15bad5dd7cb545cf61eb7b42be9a0a1fe9ea5945f04b2e23ff"} Dec 06 08:18:36 crc kubenswrapper[4945]: I1206 08:18:36.891107 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzkxz" event={"ID":"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e","Type":"ContainerStarted","Data":"a51a8c5fd8acb47bcf3b387235818493f0330f137ac84b22eea5073864a6dada"} Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.738244 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-t2htt"] Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.739608 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.742087 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fcktw" Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.742171 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.781749 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-t2htt"] Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.883701 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5c9v\" (UniqueName: \"kubernetes.io/projected/ee0e0648-81a5-4118-9e0f-53ad852f145e-kube-api-access-b5c9v\") pod \"barbican-db-sync-t2htt\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.883818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-combined-ca-bundle\") pod \"barbican-db-sync-t2htt\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.883938 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-db-sync-config-data\") pod \"barbican-db-sync-t2htt\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.903971 4945 generic.go:334] "Generic (PLEG): container finished" podID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerID="767b2cf8cde452ee0d5a8acd6cec3739b0830aa5a633afce5e6f6eac61adef99" exitCode=0 Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.904013 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzkxz" event={"ID":"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e","Type":"ContainerDied","Data":"767b2cf8cde452ee0d5a8acd6cec3739b0830aa5a633afce5e6f6eac61adef99"} Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.985738 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5c9v\" (UniqueName: \"kubernetes.io/projected/ee0e0648-81a5-4118-9e0f-53ad852f145e-kube-api-access-b5c9v\") pod \"barbican-db-sync-t2htt\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.985805 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-combined-ca-bundle\") pod \"barbican-db-sync-t2htt\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:37 crc kubenswrapper[4945]: I1206 08:18:37.985866 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-db-sync-config-data\") pod \"barbican-db-sync-t2htt\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:38 crc kubenswrapper[4945]: I1206 08:18:38.000482 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-combined-ca-bundle\") pod \"barbican-db-sync-t2htt\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:38 crc kubenswrapper[4945]: I1206 08:18:38.002718 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-db-sync-config-data\") pod \"barbican-db-sync-t2htt\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:38 crc kubenswrapper[4945]: I1206 08:18:38.003433 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5c9v\" (UniqueName: \"kubernetes.io/projected/ee0e0648-81a5-4118-9e0f-53ad852f145e-kube-api-access-b5c9v\") pod \"barbican-db-sync-t2htt\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:38 crc kubenswrapper[4945]: I1206 08:18:38.059507 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:38 crc kubenswrapper[4945]: I1206 08:18:38.476978 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-t2htt"] Dec 06 08:18:38 crc kubenswrapper[4945]: I1206 08:18:38.918859 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzkxz" event={"ID":"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e","Type":"ContainerStarted","Data":"d51438499e38543dc1fe5c2a03408fde829f3d42b5e01e1b5a7f58c05cd5ade3"} Dec 06 08:18:38 crc kubenswrapper[4945]: I1206 08:18:38.920814 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-t2htt" event={"ID":"ee0e0648-81a5-4118-9e0f-53ad852f145e","Type":"ContainerStarted","Data":"bfce7db9f8482b91f4982914b3d40fed98e60b2376f9ffd2eb81593ef53b3579"} Dec 06 08:18:38 crc kubenswrapper[4945]: I1206 08:18:38.938784 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tzkxz" podStartSLOduration=2.533892958 podStartE2EDuration="3.938769194s" podCreationTimestamp="2025-12-06 08:18:35 +0000 UTC" firstStartedPulling="2025-12-06 08:18:36.89357878 +0000 UTC m=+5130.348439824" lastFinishedPulling="2025-12-06 08:18:38.298455016 +0000 UTC m=+5131.753316060" observedRunningTime="2025-12-06 08:18:38.93751601 +0000 UTC m=+5132.392377054" watchObservedRunningTime="2025-12-06 08:18:38.938769194 +0000 UTC m=+5132.393630238" Dec 06 08:18:43 crc kubenswrapper[4945]: I1206 08:18:43.964594 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-t2htt" event={"ID":"ee0e0648-81a5-4118-9e0f-53ad852f145e","Type":"ContainerStarted","Data":"05f147cc3615e396cd747214bc8f0f527438fdca0ae3ee4612a8f1fedfc3b151"} Dec 06 08:18:43 crc kubenswrapper[4945]: I1206 08:18:43.987335 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-t2htt" podStartSLOduration=2.656527001 podStartE2EDuration="6.987310213s" podCreationTimestamp="2025-12-06 08:18:37 +0000 UTC" firstStartedPulling="2025-12-06 08:18:38.480448632 +0000 UTC m=+5131.935309666" lastFinishedPulling="2025-12-06 08:18:42.811231834 +0000 UTC m=+5136.266092878" observedRunningTime="2025-12-06 08:18:43.977600013 +0000 UTC m=+5137.432461067" watchObservedRunningTime="2025-12-06 08:18:43.987310213 +0000 UTC m=+5137.442171267" Dec 06 08:18:44 crc kubenswrapper[4945]: I1206 08:18:44.972563 4945 generic.go:334] "Generic (PLEG): container finished" podID="ee0e0648-81a5-4118-9e0f-53ad852f145e" containerID="05f147cc3615e396cd747214bc8f0f527438fdca0ae3ee4612a8f1fedfc3b151" exitCode=0 Dec 06 08:18:44 crc kubenswrapper[4945]: I1206 08:18:44.972608 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-t2htt" event={"ID":"ee0e0648-81a5-4118-9e0f-53ad852f145e","Type":"ContainerDied","Data":"05f147cc3615e396cd747214bc8f0f527438fdca0ae3ee4612a8f1fedfc3b151"} Dec 06 08:18:45 crc kubenswrapper[4945]: I1206 08:18:45.741578 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:45 crc kubenswrapper[4945]: I1206 08:18:45.741957 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:45 crc kubenswrapper[4945]: I1206 08:18:45.813011 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.032853 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.093882 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tzkxz"] Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.253205 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.325926 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5c9v\" (UniqueName: \"kubernetes.io/projected/ee0e0648-81a5-4118-9e0f-53ad852f145e-kube-api-access-b5c9v\") pod \"ee0e0648-81a5-4118-9e0f-53ad852f145e\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.326153 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-db-sync-config-data\") pod \"ee0e0648-81a5-4118-9e0f-53ad852f145e\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.326204 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-combined-ca-bundle\") pod \"ee0e0648-81a5-4118-9e0f-53ad852f145e\" (UID: \"ee0e0648-81a5-4118-9e0f-53ad852f145e\") " Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.332026 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee0e0648-81a5-4118-9e0f-53ad852f145e-kube-api-access-b5c9v" (OuterVolumeSpecName: "kube-api-access-b5c9v") pod "ee0e0648-81a5-4118-9e0f-53ad852f145e" (UID: "ee0e0648-81a5-4118-9e0f-53ad852f145e"). InnerVolumeSpecName "kube-api-access-b5c9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.333462 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ee0e0648-81a5-4118-9e0f-53ad852f145e" (UID: "ee0e0648-81a5-4118-9e0f-53ad852f145e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.348535 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee0e0648-81a5-4118-9e0f-53ad852f145e" (UID: "ee0e0648-81a5-4118-9e0f-53ad852f145e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.430580 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.430639 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee0e0648-81a5-4118-9e0f-53ad852f145e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:46 crc kubenswrapper[4945]: I1206 08:18:46.430663 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5c9v\" (UniqueName: \"kubernetes.io/projected/ee0e0648-81a5-4118-9e0f-53ad852f145e-kube-api-access-b5c9v\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.011932 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-t2htt" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.011926 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-t2htt" event={"ID":"ee0e0648-81a5-4118-9e0f-53ad852f145e","Type":"ContainerDied","Data":"bfce7db9f8482b91f4982914b3d40fed98e60b2376f9ffd2eb81593ef53b3579"} Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.011987 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfce7db9f8482b91f4982914b3d40fed98e60b2376f9ffd2eb81593ef53b3579" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.223972 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-65f7bb6fb8-2265g"] Dec 06 08:18:47 crc kubenswrapper[4945]: E1206 08:18:47.224593 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee0e0648-81a5-4118-9e0f-53ad852f145e" containerName="barbican-db-sync" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.224606 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0e0648-81a5-4118-9e0f-53ad852f145e" containerName="barbican-db-sync" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.224790 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee0e0648-81a5-4118-9e0f-53ad852f145e" containerName="barbican-db-sync" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.225673 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.228683 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.228770 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.228990 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fcktw" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.235212 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-c6cdc9cfc-2vgms"] Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.237181 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.239079 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.246210 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-65f7bb6fb8-2265g"] Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.260833 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-c6cdc9cfc-2vgms"] Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.343216 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6679b8d799-n8gmh"] Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.344556 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.348094 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9fb1fb-edb3-4bdb-9828-d24b824ee635-logs\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.348136 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430db434-86b0-48d5-a730-86fffbb325e2-combined-ca-bundle\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.348167 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be9fb1fb-edb3-4bdb-9828-d24b824ee635-config-data-custom\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.348194 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430db434-86b0-48d5-a730-86fffbb325e2-config-data\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.348226 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430db434-86b0-48d5-a730-86fffbb325e2-logs\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.348248 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9fb1fb-edb3-4bdb-9828-d24b824ee635-combined-ca-bundle\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.348305 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4z5s\" (UniqueName: \"kubernetes.io/projected/430db434-86b0-48d5-a730-86fffbb325e2-kube-api-access-q4z5s\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.348441 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm8dg\" (UniqueName: \"kubernetes.io/projected/be9fb1fb-edb3-4bdb-9828-d24b824ee635-kube-api-access-vm8dg\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.348497 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9fb1fb-edb3-4bdb-9828-d24b824ee635-config-data\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.348516 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/430db434-86b0-48d5-a730-86fffbb325e2-config-data-custom\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.359902 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6679b8d799-n8gmh"] Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.446561 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-87db876d6-6d2b6"] Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.448527 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.449990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm8dg\" (UniqueName: \"kubernetes.io/projected/be9fb1fb-edb3-4bdb-9828-d24b824ee635-kube-api-access-vm8dg\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450066 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-config\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450090 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9fb1fb-edb3-4bdb-9828-d24b824ee635-config-data\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450108 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/430db434-86b0-48d5-a730-86fffbb325e2-config-data-custom\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450688 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cstx6\" (UniqueName: \"kubernetes.io/projected/f0f86a9a-4f9b-487a-98a6-6fb733642edb-kube-api-access-cstx6\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450757 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9fb1fb-edb3-4bdb-9828-d24b824ee635-logs\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450787 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430db434-86b0-48d5-a730-86fffbb325e2-combined-ca-bundle\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450852 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be9fb1fb-edb3-4bdb-9828-d24b824ee635-config-data-custom\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450881 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430db434-86b0-48d5-a730-86fffbb325e2-config-data\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450920 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-sb\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450948 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430db434-86b0-48d5-a730-86fffbb325e2-logs\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.450969 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9fb1fb-edb3-4bdb-9828-d24b824ee635-combined-ca-bundle\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.451018 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-nb\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.451045 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4z5s\" (UniqueName: \"kubernetes.io/projected/430db434-86b0-48d5-a730-86fffbb325e2-kube-api-access-q4z5s\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.451083 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-dns-svc\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.452655 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9fb1fb-edb3-4bdb-9828-d24b824ee635-logs\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.454692 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.454796 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430db434-86b0-48d5-a730-86fffbb325e2-logs\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.465829 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9fb1fb-edb3-4bdb-9828-d24b824ee635-config-data\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.480377 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430db434-86b0-48d5-a730-86fffbb325e2-config-data\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.480856 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/430db434-86b0-48d5-a730-86fffbb325e2-config-data-custom\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.481769 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be9fb1fb-edb3-4bdb-9828-d24b824ee635-config-data-custom\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.490435 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430db434-86b0-48d5-a730-86fffbb325e2-combined-ca-bundle\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.490624 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9fb1fb-edb3-4bdb-9828-d24b824ee635-combined-ca-bundle\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.492328 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-87db876d6-6d2b6"] Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.492834 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm8dg\" (UniqueName: \"kubernetes.io/projected/be9fb1fb-edb3-4bdb-9828-d24b824ee635-kube-api-access-vm8dg\") pod \"barbican-worker-c6cdc9cfc-2vgms\" (UID: \"be9fb1fb-edb3-4bdb-9828-d24b824ee635\") " pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.493341 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4z5s\" (UniqueName: \"kubernetes.io/projected/430db434-86b0-48d5-a730-86fffbb325e2-kube-api-access-q4z5s\") pod \"barbican-keystone-listener-65f7bb6fb8-2265g\" (UID: \"430db434-86b0-48d5-a730-86fffbb325e2\") " pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.553750 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.556086 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-sb\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.556153 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73793c5e-8c22-47c2-b798-76d6f8226d73-config-data-custom\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.556220 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73793c5e-8c22-47c2-b798-76d6f8226d73-logs\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.556252 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppng6\" (UniqueName: \"kubernetes.io/projected/73793c5e-8c22-47c2-b798-76d6f8226d73-kube-api-access-ppng6\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.556308 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-nb\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.556339 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-dns-svc\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.556382 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-config\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.556419 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cstx6\" (UniqueName: \"kubernetes.io/projected/f0f86a9a-4f9b-487a-98a6-6fb733642edb-kube-api-access-cstx6\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.556453 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73793c5e-8c22-47c2-b798-76d6f8226d73-config-data\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.556526 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73793c5e-8c22-47c2-b798-76d6f8226d73-combined-ca-bundle\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.557237 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-sb\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.557490 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-nb\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.558060 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-config\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.558568 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-dns-svc\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.564693 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c6cdc9cfc-2vgms" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.601062 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cstx6\" (UniqueName: \"kubernetes.io/projected/f0f86a9a-4f9b-487a-98a6-6fb733642edb-kube-api-access-cstx6\") pod \"dnsmasq-dns-6679b8d799-n8gmh\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.657504 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73793c5e-8c22-47c2-b798-76d6f8226d73-combined-ca-bundle\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.657587 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73793c5e-8c22-47c2-b798-76d6f8226d73-config-data-custom\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.657649 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73793c5e-8c22-47c2-b798-76d6f8226d73-logs\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.657677 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppng6\" (UniqueName: \"kubernetes.io/projected/73793c5e-8c22-47c2-b798-76d6f8226d73-kube-api-access-ppng6\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.657785 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73793c5e-8c22-47c2-b798-76d6f8226d73-config-data\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.658642 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73793c5e-8c22-47c2-b798-76d6f8226d73-logs\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.660426 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.661422 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73793c5e-8c22-47c2-b798-76d6f8226d73-combined-ca-bundle\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.664815 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73793c5e-8c22-47c2-b798-76d6f8226d73-config-data-custom\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.667954 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73793c5e-8c22-47c2-b798-76d6f8226d73-config-data\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.680954 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppng6\" (UniqueName: \"kubernetes.io/projected/73793c5e-8c22-47c2-b798-76d6f8226d73-kube-api-access-ppng6\") pod \"barbican-api-87db876d6-6d2b6\" (UID: \"73793c5e-8c22-47c2-b798-76d6f8226d73\") " pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:47 crc kubenswrapper[4945]: I1206 08:18:47.866549 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:48 crc kubenswrapper[4945]: I1206 08:18:48.021257 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tzkxz" podUID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerName="registry-server" containerID="cri-o://d51438499e38543dc1fe5c2a03408fde829f3d42b5e01e1b5a7f58c05cd5ade3" gracePeriod=2 Dec 06 08:18:48 crc kubenswrapper[4945]: I1206 08:18:48.070574 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-65f7bb6fb8-2265g"] Dec 06 08:18:48 crc kubenswrapper[4945]: I1206 08:18:48.129880 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-c6cdc9cfc-2vgms"] Dec 06 08:18:48 crc kubenswrapper[4945]: W1206 08:18:48.138329 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe9fb1fb_edb3_4bdb_9828_d24b824ee635.slice/crio-61c26a32f4d34326639dc7877da270665f076784a14378205dc50e1a57f0ac90 WatchSource:0}: Error finding container 61c26a32f4d34326639dc7877da270665f076784a14378205dc50e1a57f0ac90: Status 404 returned error can't find the container with id 61c26a32f4d34326639dc7877da270665f076784a14378205dc50e1a57f0ac90 Dec 06 08:18:48 crc kubenswrapper[4945]: I1206 08:18:48.209798 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6679b8d799-n8gmh"] Dec 06 08:18:48 crc kubenswrapper[4945]: W1206 08:18:48.256781 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0f86a9a_4f9b_487a_98a6_6fb733642edb.slice/crio-83c97b9e4159a42f9b334f2e6747160b00b5c73346f315e094d64c0dad2a2550 WatchSource:0}: Error finding container 83c97b9e4159a42f9b334f2e6747160b00b5c73346f315e094d64c0dad2a2550: Status 404 returned error can't find the container with id 83c97b9e4159a42f9b334f2e6747160b00b5c73346f315e094d64c0dad2a2550 Dec 06 08:18:48 crc kubenswrapper[4945]: I1206 08:18:48.352619 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-87db876d6-6d2b6"] Dec 06 08:18:48 crc kubenswrapper[4945]: W1206 08:18:48.363480 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73793c5e_8c22_47c2_b798_76d6f8226d73.slice/crio-810db42320517407a7c0fae4668097f23733baed553547ca8e5c4d86989b6ac3 WatchSource:0}: Error finding container 810db42320517407a7c0fae4668097f23733baed553547ca8e5c4d86989b6ac3: Status 404 returned error can't find the container with id 810db42320517407a7c0fae4668097f23733baed553547ca8e5c4d86989b6ac3 Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.037709 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-87db876d6-6d2b6" event={"ID":"73793c5e-8c22-47c2-b798-76d6f8226d73","Type":"ContainerStarted","Data":"40f6cf78d52d71aa4535b2b2a53d329881773cfd7233af90d784754a0362d1e8"} Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.038135 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-87db876d6-6d2b6" event={"ID":"73793c5e-8c22-47c2-b798-76d6f8226d73","Type":"ContainerStarted","Data":"bc1bac5e078f408a799843176c2883a9c417ba5c22a0fd7c012cebfeb5205177"} Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.038155 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-87db876d6-6d2b6" event={"ID":"73793c5e-8c22-47c2-b798-76d6f8226d73","Type":"ContainerStarted","Data":"810db42320517407a7c0fae4668097f23733baed553547ca8e5c4d86989b6ac3"} Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.038202 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.038227 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.043055 4945 generic.go:334] "Generic (PLEG): container finished" podID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerID="d51438499e38543dc1fe5c2a03408fde829f3d42b5e01e1b5a7f58c05cd5ade3" exitCode=0 Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.043142 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzkxz" event={"ID":"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e","Type":"ContainerDied","Data":"d51438499e38543dc1fe5c2a03408fde829f3d42b5e01e1b5a7f58c05cd5ade3"} Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.045657 4945 generic.go:334] "Generic (PLEG): container finished" podID="f0f86a9a-4f9b-487a-98a6-6fb733642edb" containerID="a71b8fd3a1a3aec0166661f0b1a35ffdd9bfcf95696ecee1b42033448ddf9094" exitCode=0 Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.045730 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" event={"ID":"f0f86a9a-4f9b-487a-98a6-6fb733642edb","Type":"ContainerDied","Data":"a71b8fd3a1a3aec0166661f0b1a35ffdd9bfcf95696ecee1b42033448ddf9094"} Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.045757 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" event={"ID":"f0f86a9a-4f9b-487a-98a6-6fb733642edb","Type":"ContainerStarted","Data":"83c97b9e4159a42f9b334f2e6747160b00b5c73346f315e094d64c0dad2a2550"} Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.047305 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c6cdc9cfc-2vgms" event={"ID":"be9fb1fb-edb3-4bdb-9828-d24b824ee635","Type":"ContainerStarted","Data":"61c26a32f4d34326639dc7877da270665f076784a14378205dc50e1a57f0ac90"} Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.050295 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" event={"ID":"430db434-86b0-48d5-a730-86fffbb325e2","Type":"ContainerStarted","Data":"3c76a0f42eb84b7a29ea1583acdbf111798595f2d9c4bafcbac1b87a8d9e0684"} Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.062152 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-87db876d6-6d2b6" podStartSLOduration=2.062100374 podStartE2EDuration="2.062100374s" podCreationTimestamp="2025-12-06 08:18:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:18:49.060304216 +0000 UTC m=+5142.515165260" watchObservedRunningTime="2025-12-06 08:18:49.062100374 +0000 UTC m=+5142.516961418" Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.333544 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.396299 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-catalog-content\") pod \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.396717 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlr8q\" (UniqueName: \"kubernetes.io/projected/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-kube-api-access-wlr8q\") pod \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.396819 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-utilities\") pod \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\" (UID: \"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e\") " Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.398188 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-utilities" (OuterVolumeSpecName: "utilities") pod "15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" (UID: "15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.404068 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-kube-api-access-wlr8q" (OuterVolumeSpecName: "kube-api-access-wlr8q") pod "15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" (UID: "15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e"). InnerVolumeSpecName "kube-api-access-wlr8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.420369 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" (UID: "15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.498921 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlr8q\" (UniqueName: \"kubernetes.io/projected/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-kube-api-access-wlr8q\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.499115 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:49 crc kubenswrapper[4945]: I1206 08:18:49.499201 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.066781 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzkxz" event={"ID":"15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e","Type":"ContainerDied","Data":"a51a8c5fd8acb47bcf3b387235818493f0330f137ac84b22eea5073864a6dada"} Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.067130 4945 scope.go:117] "RemoveContainer" containerID="d51438499e38543dc1fe5c2a03408fde829f3d42b5e01e1b5a7f58c05cd5ade3" Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.067504 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tzkxz" Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.075570 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" event={"ID":"f0f86a9a-4f9b-487a-98a6-6fb733642edb","Type":"ContainerStarted","Data":"ababc9b1e1f5da53bf58818c8a38eb635e05c168eacca390ea6da51a37d60f2a"} Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.075864 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.078788 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c6cdc9cfc-2vgms" event={"ID":"be9fb1fb-edb3-4bdb-9828-d24b824ee635","Type":"ContainerStarted","Data":"aeed8418ff6b7e81a684245a628284af1859337290349eea986d5ce617db065c"} Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.081034 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" event={"ID":"430db434-86b0-48d5-a730-86fffbb325e2","Type":"ContainerStarted","Data":"7e7b1eec26159ff7bf040e66da964d4428990d674ada299a9e375481d0c4b951"} Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.088858 4945 scope.go:117] "RemoveContainer" containerID="767b2cf8cde452ee0d5a8acd6cec3739b0830aa5a633afce5e6f6eac61adef99" Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.106803 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" podStartSLOduration=3.106746101 podStartE2EDuration="3.106746101s" podCreationTimestamp="2025-12-06 08:18:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:18:50.093061575 +0000 UTC m=+5143.547922619" watchObservedRunningTime="2025-12-06 08:18:50.106746101 +0000 UTC m=+5143.561607145" Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.117263 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-c6cdc9cfc-2vgms" podStartSLOduration=1.6068165030000001 podStartE2EDuration="3.117246511s" podCreationTimestamp="2025-12-06 08:18:47 +0000 UTC" firstStartedPulling="2025-12-06 08:18:48.140838887 +0000 UTC m=+5141.595699931" lastFinishedPulling="2025-12-06 08:18:49.651268885 +0000 UTC m=+5143.106129939" observedRunningTime="2025-12-06 08:18:50.111069116 +0000 UTC m=+5143.565930160" watchObservedRunningTime="2025-12-06 08:18:50.117246511 +0000 UTC m=+5143.572107555" Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.125611 4945 scope.go:117] "RemoveContainer" containerID="c512cd3d350fdf15bad5dd7cb545cf61eb7b42be9a0a1fe9ea5945f04b2e23ff" Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.139670 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tzkxz"] Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.151193 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tzkxz"] Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.152127 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" podStartSLOduration=1.578702592 podStartE2EDuration="3.152114023s" podCreationTimestamp="2025-12-06 08:18:47 +0000 UTC" firstStartedPulling="2025-12-06 08:18:48.07922057 +0000 UTC m=+5141.534081624" lastFinishedPulling="2025-12-06 08:18:49.652632011 +0000 UTC m=+5143.107493055" observedRunningTime="2025-12-06 08:18:50.149168505 +0000 UTC m=+5143.604029549" watchObservedRunningTime="2025-12-06 08:18:50.152114023 +0000 UTC m=+5143.606975067" Dec 06 08:18:50 crc kubenswrapper[4945]: I1206 08:18:50.962642 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" path="/var/lib/kubelet/pods/15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e/volumes" Dec 06 08:18:51 crc kubenswrapper[4945]: I1206 08:18:51.090599 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c6cdc9cfc-2vgms" event={"ID":"be9fb1fb-edb3-4bdb-9828-d24b824ee635","Type":"ContainerStarted","Data":"3a6e46a28bbc2562cdd5accd069200d0c8f384f1c089141a355bd1f848683da0"} Dec 06 08:18:51 crc kubenswrapper[4945]: I1206 08:18:51.094254 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-65f7bb6fb8-2265g" event={"ID":"430db434-86b0-48d5-a730-86fffbb325e2","Type":"ContainerStarted","Data":"401a3d6e211a1a128cb43b5f6154f942a627a9d716dacb5868277fb503e84f7c"} Dec 06 08:18:57 crc kubenswrapper[4945]: I1206 08:18:57.667509 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:18:57 crc kubenswrapper[4945]: I1206 08:18:57.723365 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76b898c7cf-667f7"] Dec 06 08:18:57 crc kubenswrapper[4945]: I1206 08:18:57.723600 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" podUID="6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" containerName="dnsmasq-dns" containerID="cri-o://6f2fe87aec157c5fa0c110ce5891adb88be8cea215dbdf8fea66d28874545840" gracePeriod=10 Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.170637 4945 generic.go:334] "Generic (PLEG): container finished" podID="6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" containerID="6f2fe87aec157c5fa0c110ce5891adb88be8cea215dbdf8fea66d28874545840" exitCode=0 Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.171037 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" event={"ID":"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569","Type":"ContainerDied","Data":"6f2fe87aec157c5fa0c110ce5891adb88be8cea215dbdf8fea66d28874545840"} Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.269833 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.367052 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-nb\") pod \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.367166 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phk6f\" (UniqueName: \"kubernetes.io/projected/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-kube-api-access-phk6f\") pod \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.367212 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-config\") pod \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.367267 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-dns-svc\") pod \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.367340 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-sb\") pod \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\" (UID: \"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569\") " Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.373392 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-kube-api-access-phk6f" (OuterVolumeSpecName: "kube-api-access-phk6f") pod "6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" (UID: "6ffa54f0-4b4a-4c9b-a8bd-f526618ec569"). InnerVolumeSpecName "kube-api-access-phk6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.408985 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" (UID: "6ffa54f0-4b4a-4c9b-a8bd-f526618ec569"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.417110 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-config" (OuterVolumeSpecName: "config") pod "6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" (UID: "6ffa54f0-4b4a-4c9b-a8bd-f526618ec569"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.428205 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" (UID: "6ffa54f0-4b4a-4c9b-a8bd-f526618ec569"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.428971 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" (UID: "6ffa54f0-4b4a-4c9b-a8bd-f526618ec569"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.469629 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phk6f\" (UniqueName: \"kubernetes.io/projected/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-kube-api-access-phk6f\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.469659 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.469669 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.469680 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:58 crc kubenswrapper[4945]: I1206 08:18:58.469688 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:18:59 crc kubenswrapper[4945]: I1206 08:18:59.182256 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" event={"ID":"6ffa54f0-4b4a-4c9b-a8bd-f526618ec569","Type":"ContainerDied","Data":"8715f134363d1a8badf921f020d7c24efaec89e91c70aa7fb32d5405f41c91d5"} Dec 06 08:18:59 crc kubenswrapper[4945]: I1206 08:18:59.182692 4945 scope.go:117] "RemoveContainer" containerID="6f2fe87aec157c5fa0c110ce5891adb88be8cea215dbdf8fea66d28874545840" Dec 06 08:18:59 crc kubenswrapper[4945]: I1206 08:18:59.182449 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" Dec 06 08:18:59 crc kubenswrapper[4945]: I1206 08:18:59.211557 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76b898c7cf-667f7"] Dec 06 08:18:59 crc kubenswrapper[4945]: I1206 08:18:59.217400 4945 scope.go:117] "RemoveContainer" containerID="5493a13249c18257ae4481e5af502376ea5cf825e31781b3fdeed683101c4acf" Dec 06 08:18:59 crc kubenswrapper[4945]: I1206 08:18:59.220213 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76b898c7cf-667f7"] Dec 06 08:18:59 crc kubenswrapper[4945]: I1206 08:18:59.320683 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:18:59 crc kubenswrapper[4945]: I1206 08:18:59.502804 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-87db876d6-6d2b6" Dec 06 08:19:00 crc kubenswrapper[4945]: I1206 08:19:00.975257 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" path="/var/lib/kubelet/pods/6ffa54f0-4b4a-4c9b-a8bd-f526618ec569/volumes" Dec 06 08:19:03 crc kubenswrapper[4945]: I1206 08:19:03.149183 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76b898c7cf-667f7" podUID="6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.16:5353: i/o timeout" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.404742 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hflgw"] Dec 06 08:19:07 crc kubenswrapper[4945]: E1206 08:19:07.407325 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerName="extract-utilities" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.407351 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerName="extract-utilities" Dec 06 08:19:07 crc kubenswrapper[4945]: E1206 08:19:07.407365 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" containerName="init" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.407372 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" containerName="init" Dec 06 08:19:07 crc kubenswrapper[4945]: E1206 08:19:07.407394 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerName="registry-server" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.407401 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerName="registry-server" Dec 06 08:19:07 crc kubenswrapper[4945]: E1206 08:19:07.407418 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerName="extract-content" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.407425 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerName="extract-content" Dec 06 08:19:07 crc kubenswrapper[4945]: E1206 08:19:07.407439 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" containerName="dnsmasq-dns" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.407445 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" containerName="dnsmasq-dns" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.407899 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ffa54f0-4b4a-4c9b-a8bd-f526618ec569" containerName="dnsmasq-dns" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.407918 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="15f29b5a-e0ef-4265-a2dc-8f159e5a8a7e" containerName="registry-server" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.413212 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.443259 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hflgw"] Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.536015 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-catalog-content\") pod \"certified-operators-hflgw\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.536367 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2bxj\" (UniqueName: \"kubernetes.io/projected/50ea82ed-7905-409a-a3c9-b723bc7c2d59-kube-api-access-q2bxj\") pod \"certified-operators-hflgw\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.536435 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-utilities\") pod \"certified-operators-hflgw\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.637801 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-catalog-content\") pod \"certified-operators-hflgw\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.637947 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2bxj\" (UniqueName: \"kubernetes.io/projected/50ea82ed-7905-409a-a3c9-b723bc7c2d59-kube-api-access-q2bxj\") pod \"certified-operators-hflgw\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.637971 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-utilities\") pod \"certified-operators-hflgw\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.638300 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-catalog-content\") pod \"certified-operators-hflgw\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.638511 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-utilities\") pod \"certified-operators-hflgw\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.661845 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2bxj\" (UniqueName: \"kubernetes.io/projected/50ea82ed-7905-409a-a3c9-b723bc7c2d59-kube-api-access-q2bxj\") pod \"certified-operators-hflgw\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:07 crc kubenswrapper[4945]: I1206 08:19:07.775833 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:08 crc kubenswrapper[4945]: I1206 08:19:08.300831 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hflgw"] Dec 06 08:19:08 crc kubenswrapper[4945]: W1206 08:19:08.302547 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50ea82ed_7905_409a_a3c9_b723bc7c2d59.slice/crio-06a71d9dc1df7005327a6d13495f9b296988008ad3c50241f6afacbeaf88307f WatchSource:0}: Error finding container 06a71d9dc1df7005327a6d13495f9b296988008ad3c50241f6afacbeaf88307f: Status 404 returned error can't find the container with id 06a71d9dc1df7005327a6d13495f9b296988008ad3c50241f6afacbeaf88307f Dec 06 08:19:09 crc kubenswrapper[4945]: I1206 08:19:09.275903 4945 generic.go:334] "Generic (PLEG): container finished" podID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerID="e4a393d2d2f56bae3ffab0570ab4e4d71044556193d9631a8daf7dca842b3953" exitCode=0 Dec 06 08:19:09 crc kubenswrapper[4945]: I1206 08:19:09.275966 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hflgw" event={"ID":"50ea82ed-7905-409a-a3c9-b723bc7c2d59","Type":"ContainerDied","Data":"e4a393d2d2f56bae3ffab0570ab4e4d71044556193d9631a8daf7dca842b3953"} Dec 06 08:19:09 crc kubenswrapper[4945]: I1206 08:19:09.276176 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hflgw" event={"ID":"50ea82ed-7905-409a-a3c9-b723bc7c2d59","Type":"ContainerStarted","Data":"06a71d9dc1df7005327a6d13495f9b296988008ad3c50241f6afacbeaf88307f"} Dec 06 08:19:10 crc kubenswrapper[4945]: I1206 08:19:10.287416 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hflgw" event={"ID":"50ea82ed-7905-409a-a3c9-b723bc7c2d59","Type":"ContainerStarted","Data":"292a3924c781e770a3e69e8f773d55b0dce743734b3f7020c8e4e81404ca5977"} Dec 06 08:19:11 crc kubenswrapper[4945]: I1206 08:19:11.297457 4945 generic.go:334] "Generic (PLEG): container finished" podID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerID="292a3924c781e770a3e69e8f773d55b0dce743734b3f7020c8e4e81404ca5977" exitCode=0 Dec 06 08:19:11 crc kubenswrapper[4945]: I1206 08:19:11.297494 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hflgw" event={"ID":"50ea82ed-7905-409a-a3c9-b723bc7c2d59","Type":"ContainerDied","Data":"292a3924c781e770a3e69e8f773d55b0dce743734b3f7020c8e4e81404ca5977"} Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.306386 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hflgw" event={"ID":"50ea82ed-7905-409a-a3c9-b723bc7c2d59","Type":"ContainerStarted","Data":"2ed34782d60ba4933cd8eaaca4d88ec1e6b02cebdaad532d32764a5c27b12537"} Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.325358 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hflgw" podStartSLOduration=2.8756048400000003 podStartE2EDuration="5.325338728s" podCreationTimestamp="2025-12-06 08:19:07 +0000 UTC" firstStartedPulling="2025-12-06 08:19:09.277751279 +0000 UTC m=+5162.732612323" lastFinishedPulling="2025-12-06 08:19:11.727485167 +0000 UTC m=+5165.182346211" observedRunningTime="2025-12-06 08:19:12.323551 +0000 UTC m=+5165.778412064" watchObservedRunningTime="2025-12-06 08:19:12.325338728 +0000 UTC m=+5165.780199782" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.417524 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-tpdn5"] Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.418914 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tpdn5" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.428645 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tpdn5"] Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.513620 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7c24-account-create-update-cgl54"] Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.514556 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhs9t\" (UniqueName: \"kubernetes.io/projected/e9a17984-24af-49d1-ad27-54655f405ee8-kube-api-access-mhs9t\") pod \"neutron-db-create-tpdn5\" (UID: \"e9a17984-24af-49d1-ad27-54655f405ee8\") " pod="openstack/neutron-db-create-tpdn5" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.514678 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e9a17984-24af-49d1-ad27-54655f405ee8-operator-scripts\") pod \"neutron-db-create-tpdn5\" (UID: \"e9a17984-24af-49d1-ad27-54655f405ee8\") " pod="openstack/neutron-db-create-tpdn5" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.514744 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c24-account-create-update-cgl54" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.519689 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.524117 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c24-account-create-update-cgl54"] Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.616227 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b76d4-f78c-4972-832c-c586920f6a0f-operator-scripts\") pod \"neutron-7c24-account-create-update-cgl54\" (UID: \"5a1b76d4-f78c-4972-832c-c586920f6a0f\") " pod="openstack/neutron-7c24-account-create-update-cgl54" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.616333 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhs9t\" (UniqueName: \"kubernetes.io/projected/e9a17984-24af-49d1-ad27-54655f405ee8-kube-api-access-mhs9t\") pod \"neutron-db-create-tpdn5\" (UID: \"e9a17984-24af-49d1-ad27-54655f405ee8\") " pod="openstack/neutron-db-create-tpdn5" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.616463 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e9a17984-24af-49d1-ad27-54655f405ee8-operator-scripts\") pod \"neutron-db-create-tpdn5\" (UID: \"e9a17984-24af-49d1-ad27-54655f405ee8\") " pod="openstack/neutron-db-create-tpdn5" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.616490 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q79gz\" (UniqueName: \"kubernetes.io/projected/5a1b76d4-f78c-4972-832c-c586920f6a0f-kube-api-access-q79gz\") pod \"neutron-7c24-account-create-update-cgl54\" (UID: \"5a1b76d4-f78c-4972-832c-c586920f6a0f\") " pod="openstack/neutron-7c24-account-create-update-cgl54" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.617420 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e9a17984-24af-49d1-ad27-54655f405ee8-operator-scripts\") pod \"neutron-db-create-tpdn5\" (UID: \"e9a17984-24af-49d1-ad27-54655f405ee8\") " pod="openstack/neutron-db-create-tpdn5" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.639027 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhs9t\" (UniqueName: \"kubernetes.io/projected/e9a17984-24af-49d1-ad27-54655f405ee8-kube-api-access-mhs9t\") pod \"neutron-db-create-tpdn5\" (UID: \"e9a17984-24af-49d1-ad27-54655f405ee8\") " pod="openstack/neutron-db-create-tpdn5" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.717445 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q79gz\" (UniqueName: \"kubernetes.io/projected/5a1b76d4-f78c-4972-832c-c586920f6a0f-kube-api-access-q79gz\") pod \"neutron-7c24-account-create-update-cgl54\" (UID: \"5a1b76d4-f78c-4972-832c-c586920f6a0f\") " pod="openstack/neutron-7c24-account-create-update-cgl54" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.717515 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b76d4-f78c-4972-832c-c586920f6a0f-operator-scripts\") pod \"neutron-7c24-account-create-update-cgl54\" (UID: \"5a1b76d4-f78c-4972-832c-c586920f6a0f\") " pod="openstack/neutron-7c24-account-create-update-cgl54" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.718338 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b76d4-f78c-4972-832c-c586920f6a0f-operator-scripts\") pod \"neutron-7c24-account-create-update-cgl54\" (UID: \"5a1b76d4-f78c-4972-832c-c586920f6a0f\") " pod="openstack/neutron-7c24-account-create-update-cgl54" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.737897 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q79gz\" (UniqueName: \"kubernetes.io/projected/5a1b76d4-f78c-4972-832c-c586920f6a0f-kube-api-access-q79gz\") pod \"neutron-7c24-account-create-update-cgl54\" (UID: \"5a1b76d4-f78c-4972-832c-c586920f6a0f\") " pod="openstack/neutron-7c24-account-create-update-cgl54" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.747102 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tpdn5" Dec 06 08:19:12 crc kubenswrapper[4945]: I1206 08:19:12.829666 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c24-account-create-update-cgl54" Dec 06 08:19:13 crc kubenswrapper[4945]: I1206 08:19:13.202380 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tpdn5"] Dec 06 08:19:13 crc kubenswrapper[4945]: I1206 08:19:13.317499 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tpdn5" event={"ID":"e9a17984-24af-49d1-ad27-54655f405ee8","Type":"ContainerStarted","Data":"65674a399a1763c46d384c01a26f6b5476dd0475f9f17c14cdb82286128179fa"} Dec 06 08:19:13 crc kubenswrapper[4945]: I1206 08:19:13.323040 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c24-account-create-update-cgl54"] Dec 06 08:19:13 crc kubenswrapper[4945]: W1206 08:19:13.326358 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a1b76d4_f78c_4972_832c_c586920f6a0f.slice/crio-1f3f0ea13a725853932645c9bbc301d6bd911484fcd582577d6dcbf96595dab1 WatchSource:0}: Error finding container 1f3f0ea13a725853932645c9bbc301d6bd911484fcd582577d6dcbf96595dab1: Status 404 returned error can't find the container with id 1f3f0ea13a725853932645c9bbc301d6bd911484fcd582577d6dcbf96595dab1 Dec 06 08:19:14 crc kubenswrapper[4945]: I1206 08:19:14.329440 4945 generic.go:334] "Generic (PLEG): container finished" podID="5a1b76d4-f78c-4972-832c-c586920f6a0f" containerID="fd3b89d05d6b2916c2c89ae1e0a0999380306c5d4791b2531a52dc69babb7ade" exitCode=0 Dec 06 08:19:14 crc kubenswrapper[4945]: I1206 08:19:14.329498 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c24-account-create-update-cgl54" event={"ID":"5a1b76d4-f78c-4972-832c-c586920f6a0f","Type":"ContainerDied","Data":"fd3b89d05d6b2916c2c89ae1e0a0999380306c5d4791b2531a52dc69babb7ade"} Dec 06 08:19:14 crc kubenswrapper[4945]: I1206 08:19:14.329743 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c24-account-create-update-cgl54" event={"ID":"5a1b76d4-f78c-4972-832c-c586920f6a0f","Type":"ContainerStarted","Data":"1f3f0ea13a725853932645c9bbc301d6bd911484fcd582577d6dcbf96595dab1"} Dec 06 08:19:14 crc kubenswrapper[4945]: I1206 08:19:14.332353 4945 generic.go:334] "Generic (PLEG): container finished" podID="e9a17984-24af-49d1-ad27-54655f405ee8" containerID="0362c706459ad07def527950a4c7330ed2c8ae1142960517ac2d91c820c4ec43" exitCode=0 Dec 06 08:19:14 crc kubenswrapper[4945]: I1206 08:19:14.332442 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tpdn5" event={"ID":"e9a17984-24af-49d1-ad27-54655f405ee8","Type":"ContainerDied","Data":"0362c706459ad07def527950a4c7330ed2c8ae1142960517ac2d91c820c4ec43"} Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.815211 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tpdn5" Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.818545 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c24-account-create-update-cgl54" Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.877134 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e9a17984-24af-49d1-ad27-54655f405ee8-operator-scripts\") pod \"e9a17984-24af-49d1-ad27-54655f405ee8\" (UID: \"e9a17984-24af-49d1-ad27-54655f405ee8\") " Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.879129 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9a17984-24af-49d1-ad27-54655f405ee8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e9a17984-24af-49d1-ad27-54655f405ee8" (UID: "e9a17984-24af-49d1-ad27-54655f405ee8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.879259 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b76d4-f78c-4972-832c-c586920f6a0f-operator-scripts\") pod \"5a1b76d4-f78c-4972-832c-c586920f6a0f\" (UID: \"5a1b76d4-f78c-4972-832c-c586920f6a0f\") " Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.879856 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1b76d4-f78c-4972-832c-c586920f6a0f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a1b76d4-f78c-4972-832c-c586920f6a0f" (UID: "5a1b76d4-f78c-4972-832c-c586920f6a0f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.882000 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q79gz\" (UniqueName: \"kubernetes.io/projected/5a1b76d4-f78c-4972-832c-c586920f6a0f-kube-api-access-q79gz\") pod \"5a1b76d4-f78c-4972-832c-c586920f6a0f\" (UID: \"5a1b76d4-f78c-4972-832c-c586920f6a0f\") " Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.882147 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhs9t\" (UniqueName: \"kubernetes.io/projected/e9a17984-24af-49d1-ad27-54655f405ee8-kube-api-access-mhs9t\") pod \"e9a17984-24af-49d1-ad27-54655f405ee8\" (UID: \"e9a17984-24af-49d1-ad27-54655f405ee8\") " Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.883988 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e9a17984-24af-49d1-ad27-54655f405ee8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.884012 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b76d4-f78c-4972-832c-c586920f6a0f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.895594 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9a17984-24af-49d1-ad27-54655f405ee8-kube-api-access-mhs9t" (OuterVolumeSpecName: "kube-api-access-mhs9t") pod "e9a17984-24af-49d1-ad27-54655f405ee8" (UID: "e9a17984-24af-49d1-ad27-54655f405ee8"). InnerVolumeSpecName "kube-api-access-mhs9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.895898 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a1b76d4-f78c-4972-832c-c586920f6a0f-kube-api-access-q79gz" (OuterVolumeSpecName: "kube-api-access-q79gz") pod "5a1b76d4-f78c-4972-832c-c586920f6a0f" (UID: "5a1b76d4-f78c-4972-832c-c586920f6a0f"). InnerVolumeSpecName "kube-api-access-q79gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.986101 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q79gz\" (UniqueName: \"kubernetes.io/projected/5a1b76d4-f78c-4972-832c-c586920f6a0f-kube-api-access-q79gz\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:15 crc kubenswrapper[4945]: I1206 08:19:15.986144 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhs9t\" (UniqueName: \"kubernetes.io/projected/e9a17984-24af-49d1-ad27-54655f405ee8-kube-api-access-mhs9t\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:16 crc kubenswrapper[4945]: I1206 08:19:16.351117 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tpdn5" Dec 06 08:19:16 crc kubenswrapper[4945]: I1206 08:19:16.351153 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tpdn5" event={"ID":"e9a17984-24af-49d1-ad27-54655f405ee8","Type":"ContainerDied","Data":"65674a399a1763c46d384c01a26f6b5476dd0475f9f17c14cdb82286128179fa"} Dec 06 08:19:16 crc kubenswrapper[4945]: I1206 08:19:16.351201 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65674a399a1763c46d384c01a26f6b5476dd0475f9f17c14cdb82286128179fa" Dec 06 08:19:16 crc kubenswrapper[4945]: I1206 08:19:16.352659 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c24-account-create-update-cgl54" event={"ID":"5a1b76d4-f78c-4972-832c-c586920f6a0f","Type":"ContainerDied","Data":"1f3f0ea13a725853932645c9bbc301d6bd911484fcd582577d6dcbf96595dab1"} Dec 06 08:19:16 crc kubenswrapper[4945]: I1206 08:19:16.352688 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f3f0ea13a725853932645c9bbc301d6bd911484fcd582577d6dcbf96595dab1" Dec 06 08:19:16 crc kubenswrapper[4945]: I1206 08:19:16.352750 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c24-account-create-update-cgl54" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.776741 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.777160 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.830661 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-g85nz"] Dec 06 08:19:17 crc kubenswrapper[4945]: E1206 08:19:17.831061 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1b76d4-f78c-4972-832c-c586920f6a0f" containerName="mariadb-account-create-update" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.831080 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1b76d4-f78c-4972-832c-c586920f6a0f" containerName="mariadb-account-create-update" Dec 06 08:19:17 crc kubenswrapper[4945]: E1206 08:19:17.831094 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a17984-24af-49d1-ad27-54655f405ee8" containerName="mariadb-database-create" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.831102 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a17984-24af-49d1-ad27-54655f405ee8" containerName="mariadb-database-create" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.831335 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a17984-24af-49d1-ad27-54655f405ee8" containerName="mariadb-database-create" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.831356 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1b76d4-f78c-4972-832c-c586920f6a0f" containerName="mariadb-account-create-update" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.832016 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.832822 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.838126 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.838628 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.838758 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-7b5k8" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.841178 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-g85nz"] Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.915379 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb5zp\" (UniqueName: \"kubernetes.io/projected/231b069e-3731-42fa-9431-1b9081ea9d0d-kube-api-access-gb5zp\") pod \"neutron-db-sync-g85nz\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.915452 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-combined-ca-bundle\") pod \"neutron-db-sync-g85nz\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:17 crc kubenswrapper[4945]: I1206 08:19:17.915511 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-config\") pod \"neutron-db-sync-g85nz\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:18 crc kubenswrapper[4945]: I1206 08:19:18.016747 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb5zp\" (UniqueName: \"kubernetes.io/projected/231b069e-3731-42fa-9431-1b9081ea9d0d-kube-api-access-gb5zp\") pod \"neutron-db-sync-g85nz\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:18 crc kubenswrapper[4945]: I1206 08:19:18.016854 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-combined-ca-bundle\") pod \"neutron-db-sync-g85nz\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:18 crc kubenswrapper[4945]: I1206 08:19:18.016908 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-config\") pod \"neutron-db-sync-g85nz\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:18 crc kubenswrapper[4945]: I1206 08:19:18.029728 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-combined-ca-bundle\") pod \"neutron-db-sync-g85nz\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:18 crc kubenswrapper[4945]: I1206 08:19:18.029809 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-config\") pod \"neutron-db-sync-g85nz\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:18 crc kubenswrapper[4945]: I1206 08:19:18.036158 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb5zp\" (UniqueName: \"kubernetes.io/projected/231b069e-3731-42fa-9431-1b9081ea9d0d-kube-api-access-gb5zp\") pod \"neutron-db-sync-g85nz\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:18 crc kubenswrapper[4945]: I1206 08:19:18.200628 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:18 crc kubenswrapper[4945]: I1206 08:19:18.428354 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:18 crc kubenswrapper[4945]: I1206 08:19:18.484572 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hflgw"] Dec 06 08:19:18 crc kubenswrapper[4945]: I1206 08:19:18.648795 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-g85nz"] Dec 06 08:19:19 crc kubenswrapper[4945]: I1206 08:19:19.382389 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g85nz" event={"ID":"231b069e-3731-42fa-9431-1b9081ea9d0d","Type":"ContainerStarted","Data":"1274ff27990c32dfd585c75d16dfab0e5d2a1d36f12b57aa1883e053c33dfc54"} Dec 06 08:19:19 crc kubenswrapper[4945]: I1206 08:19:19.382720 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g85nz" event={"ID":"231b069e-3731-42fa-9431-1b9081ea9d0d","Type":"ContainerStarted","Data":"fa9be155b6ef258cd17382b5dabafedd73548234ed39c090c7d87b3b47e69f56"} Dec 06 08:19:19 crc kubenswrapper[4945]: I1206 08:19:19.408841 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-g85nz" podStartSLOduration=2.408818807 podStartE2EDuration="2.408818807s" podCreationTimestamp="2025-12-06 08:19:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:19:19.401858551 +0000 UTC m=+5172.856719595" watchObservedRunningTime="2025-12-06 08:19:19.408818807 +0000 UTC m=+5172.863679851" Dec 06 08:19:20 crc kubenswrapper[4945]: I1206 08:19:20.388924 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hflgw" podUID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerName="registry-server" containerID="cri-o://2ed34782d60ba4933cd8eaaca4d88ec1e6b02cebdaad532d32764a5c27b12537" gracePeriod=2 Dec 06 08:19:21 crc kubenswrapper[4945]: I1206 08:19:21.401520 4945 generic.go:334] "Generic (PLEG): container finished" podID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerID="2ed34782d60ba4933cd8eaaca4d88ec1e6b02cebdaad532d32764a5c27b12537" exitCode=0 Dec 06 08:19:21 crc kubenswrapper[4945]: I1206 08:19:21.401742 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hflgw" event={"ID":"50ea82ed-7905-409a-a3c9-b723bc7c2d59","Type":"ContainerDied","Data":"2ed34782d60ba4933cd8eaaca4d88ec1e6b02cebdaad532d32764a5c27b12537"} Dec 06 08:19:21 crc kubenswrapper[4945]: I1206 08:19:21.971149 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.085927 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2bxj\" (UniqueName: \"kubernetes.io/projected/50ea82ed-7905-409a-a3c9-b723bc7c2d59-kube-api-access-q2bxj\") pod \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.086039 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-utilities\") pod \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.086078 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-catalog-content\") pod \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\" (UID: \"50ea82ed-7905-409a-a3c9-b723bc7c2d59\") " Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.087129 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-utilities" (OuterVolumeSpecName: "utilities") pod "50ea82ed-7905-409a-a3c9-b723bc7c2d59" (UID: "50ea82ed-7905-409a-a3c9-b723bc7c2d59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.091526 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50ea82ed-7905-409a-a3c9-b723bc7c2d59-kube-api-access-q2bxj" (OuterVolumeSpecName: "kube-api-access-q2bxj") pod "50ea82ed-7905-409a-a3c9-b723bc7c2d59" (UID: "50ea82ed-7905-409a-a3c9-b723bc7c2d59"). InnerVolumeSpecName "kube-api-access-q2bxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.132514 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50ea82ed-7905-409a-a3c9-b723bc7c2d59" (UID: "50ea82ed-7905-409a-a3c9-b723bc7c2d59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.187713 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2bxj\" (UniqueName: \"kubernetes.io/projected/50ea82ed-7905-409a-a3c9-b723bc7c2d59-kube-api-access-q2bxj\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.187755 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.187768 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ea82ed-7905-409a-a3c9-b723bc7c2d59-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.413264 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hflgw" event={"ID":"50ea82ed-7905-409a-a3c9-b723bc7c2d59","Type":"ContainerDied","Data":"06a71d9dc1df7005327a6d13495f9b296988008ad3c50241f6afacbeaf88307f"} Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.413362 4945 scope.go:117] "RemoveContainer" containerID="2ed34782d60ba4933cd8eaaca4d88ec1e6b02cebdaad532d32764a5c27b12537" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.413381 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hflgw" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.439141 4945 scope.go:117] "RemoveContainer" containerID="292a3924c781e770a3e69e8f773d55b0dce743734b3f7020c8e4e81404ca5977" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.447957 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hflgw"] Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.457172 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hflgw"] Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.465946 4945 scope.go:117] "RemoveContainer" containerID="e4a393d2d2f56bae3ffab0570ab4e4d71044556193d9631a8daf7dca842b3953" Dec 06 08:19:22 crc kubenswrapper[4945]: I1206 08:19:22.963875 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" path="/var/lib/kubelet/pods/50ea82ed-7905-409a-a3c9-b723bc7c2d59/volumes" Dec 06 08:19:23 crc kubenswrapper[4945]: I1206 08:19:23.422501 4945 generic.go:334] "Generic (PLEG): container finished" podID="231b069e-3731-42fa-9431-1b9081ea9d0d" containerID="1274ff27990c32dfd585c75d16dfab0e5d2a1d36f12b57aa1883e053c33dfc54" exitCode=0 Dec 06 08:19:23 crc kubenswrapper[4945]: I1206 08:19:23.422558 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g85nz" event={"ID":"231b069e-3731-42fa-9431-1b9081ea9d0d","Type":"ContainerDied","Data":"1274ff27990c32dfd585c75d16dfab0e5d2a1d36f12b57aa1883e053c33dfc54"} Dec 06 08:19:24 crc kubenswrapper[4945]: I1206 08:19:24.768342 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:24 crc kubenswrapper[4945]: I1206 08:19:24.832070 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb5zp\" (UniqueName: \"kubernetes.io/projected/231b069e-3731-42fa-9431-1b9081ea9d0d-kube-api-access-gb5zp\") pod \"231b069e-3731-42fa-9431-1b9081ea9d0d\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " Dec 06 08:19:24 crc kubenswrapper[4945]: I1206 08:19:24.832195 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-combined-ca-bundle\") pod \"231b069e-3731-42fa-9431-1b9081ea9d0d\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " Dec 06 08:19:24 crc kubenswrapper[4945]: I1206 08:19:24.832225 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-config\") pod \"231b069e-3731-42fa-9431-1b9081ea9d0d\" (UID: \"231b069e-3731-42fa-9431-1b9081ea9d0d\") " Dec 06 08:19:24 crc kubenswrapper[4945]: I1206 08:19:24.836779 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/231b069e-3731-42fa-9431-1b9081ea9d0d-kube-api-access-gb5zp" (OuterVolumeSpecName: "kube-api-access-gb5zp") pod "231b069e-3731-42fa-9431-1b9081ea9d0d" (UID: "231b069e-3731-42fa-9431-1b9081ea9d0d"). InnerVolumeSpecName "kube-api-access-gb5zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:19:24 crc kubenswrapper[4945]: I1206 08:19:24.857772 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-config" (OuterVolumeSpecName: "config") pod "231b069e-3731-42fa-9431-1b9081ea9d0d" (UID: "231b069e-3731-42fa-9431-1b9081ea9d0d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:19:24 crc kubenswrapper[4945]: I1206 08:19:24.860144 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "231b069e-3731-42fa-9431-1b9081ea9d0d" (UID: "231b069e-3731-42fa-9431-1b9081ea9d0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:19:24 crc kubenswrapper[4945]: I1206 08:19:24.934817 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb5zp\" (UniqueName: \"kubernetes.io/projected/231b069e-3731-42fa-9431-1b9081ea9d0d-kube-api-access-gb5zp\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:24 crc kubenswrapper[4945]: I1206 08:19:24.934867 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:24 crc kubenswrapper[4945]: I1206 08:19:24.934878 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/231b069e-3731-42fa-9431-1b9081ea9d0d-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.439029 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g85nz" event={"ID":"231b069e-3731-42fa-9431-1b9081ea9d0d","Type":"ContainerDied","Data":"fa9be155b6ef258cd17382b5dabafedd73548234ed39c090c7d87b3b47e69f56"} Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.439070 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa9be155b6ef258cd17382b5dabafedd73548234ed39c090c7d87b3b47e69f56" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.439099 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g85nz" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.611517 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64dd684579-jss4d"] Dec 06 08:19:25 crc kubenswrapper[4945]: E1206 08:19:25.611922 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerName="registry-server" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.611934 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerName="registry-server" Dec 06 08:19:25 crc kubenswrapper[4945]: E1206 08:19:25.611951 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerName="extract-content" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.611960 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerName="extract-content" Dec 06 08:19:25 crc kubenswrapper[4945]: E1206 08:19:25.611978 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerName="extract-utilities" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.611985 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerName="extract-utilities" Dec 06 08:19:25 crc kubenswrapper[4945]: E1206 08:19:25.612006 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="231b069e-3731-42fa-9431-1b9081ea9d0d" containerName="neutron-db-sync" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.612012 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="231b069e-3731-42fa-9431-1b9081ea9d0d" containerName="neutron-db-sync" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.612176 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="50ea82ed-7905-409a-a3c9-b723bc7c2d59" containerName="registry-server" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.612201 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="231b069e-3731-42fa-9431-1b9081ea9d0d" containerName="neutron-db-sync" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.613082 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.627828 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64dd684579-jss4d"] Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.646708 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-dns-svc\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.646821 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-config\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.646856 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29ptr\" (UniqueName: \"kubernetes.io/projected/63107fb2-24b7-4722-9e51-eae2ded28727-kube-api-access-29ptr\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.646876 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-sb\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.646925 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-nb\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.748806 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-config\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.748876 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29ptr\" (UniqueName: \"kubernetes.io/projected/63107fb2-24b7-4722-9e51-eae2ded28727-kube-api-access-29ptr\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.748904 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-sb\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.748959 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-nb\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.748991 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-dns-svc\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.749769 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-config\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.749872 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-sb\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.750080 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-dns-svc\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.750130 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-nb\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.763858 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d89777897-g2pzk"] Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.765656 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.767382 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.767540 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-7b5k8" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.770869 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.774944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29ptr\" (UniqueName: \"kubernetes.io/projected/63107fb2-24b7-4722-9e51-eae2ded28727-kube-api-access-29ptr\") pod \"dnsmasq-dns-64dd684579-jss4d\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.784955 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d89777897-g2pzk"] Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.850939 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bb80e7-5bad-496c-a404-04f4034de941-combined-ca-bundle\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.851031 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhsp7\" (UniqueName: \"kubernetes.io/projected/17bb80e7-5bad-496c-a404-04f4034de941-kube-api-access-dhsp7\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.851084 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/17bb80e7-5bad-496c-a404-04f4034de941-config\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.851103 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/17bb80e7-5bad-496c-a404-04f4034de941-httpd-config\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.939407 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.952237 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/17bb80e7-5bad-496c-a404-04f4034de941-config\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.952310 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/17bb80e7-5bad-496c-a404-04f4034de941-httpd-config\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.952489 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bb80e7-5bad-496c-a404-04f4034de941-combined-ca-bundle\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.952553 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhsp7\" (UniqueName: \"kubernetes.io/projected/17bb80e7-5bad-496c-a404-04f4034de941-kube-api-access-dhsp7\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.958656 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/17bb80e7-5bad-496c-a404-04f4034de941-config\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.958884 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/17bb80e7-5bad-496c-a404-04f4034de941-httpd-config\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.959369 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bb80e7-5bad-496c-a404-04f4034de941-combined-ca-bundle\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:25 crc kubenswrapper[4945]: I1206 08:19:25.987960 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhsp7\" (UniqueName: \"kubernetes.io/projected/17bb80e7-5bad-496c-a404-04f4034de941-kube-api-access-dhsp7\") pod \"neutron-5d89777897-g2pzk\" (UID: \"17bb80e7-5bad-496c-a404-04f4034de941\") " pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:26 crc kubenswrapper[4945]: I1206 08:19:26.131953 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:26 crc kubenswrapper[4945]: I1206 08:19:26.410007 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64dd684579-jss4d"] Dec 06 08:19:26 crc kubenswrapper[4945]: I1206 08:19:26.447033 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dd684579-jss4d" event={"ID":"63107fb2-24b7-4722-9e51-eae2ded28727","Type":"ContainerStarted","Data":"54b8d15a6d1df2116345ef5c898f2bd8df3ad028850a9b173f5af722c32948dc"} Dec 06 08:19:26 crc kubenswrapper[4945]: I1206 08:19:26.776143 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d89777897-g2pzk"] Dec 06 08:19:26 crc kubenswrapper[4945]: W1206 08:19:26.803881 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17bb80e7_5bad_496c_a404_04f4034de941.slice/crio-84c310d84600d910890891b74cf21aeb7cac26adffcafdf7315fd36fae264c33 WatchSource:0}: Error finding container 84c310d84600d910890891b74cf21aeb7cac26adffcafdf7315fd36fae264c33: Status 404 returned error can't find the container with id 84c310d84600d910890891b74cf21aeb7cac26adffcafdf7315fd36fae264c33 Dec 06 08:19:27 crc kubenswrapper[4945]: I1206 08:19:27.457811 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d89777897-g2pzk" event={"ID":"17bb80e7-5bad-496c-a404-04f4034de941","Type":"ContainerStarted","Data":"cb2b0da04625ed11e21aa153570c51beb409591647cc888276c7d84ea997fd38"} Dec 06 08:19:27 crc kubenswrapper[4945]: I1206 08:19:27.458194 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:19:27 crc kubenswrapper[4945]: I1206 08:19:27.458237 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d89777897-g2pzk" event={"ID":"17bb80e7-5bad-496c-a404-04f4034de941","Type":"ContainerStarted","Data":"7c540cfb0b881d501ff817d05eb52a50db9671ab2e57c78e4211139009727b52"} Dec 06 08:19:27 crc kubenswrapper[4945]: I1206 08:19:27.458264 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d89777897-g2pzk" event={"ID":"17bb80e7-5bad-496c-a404-04f4034de941","Type":"ContainerStarted","Data":"84c310d84600d910890891b74cf21aeb7cac26adffcafdf7315fd36fae264c33"} Dec 06 08:19:27 crc kubenswrapper[4945]: I1206 08:19:27.459501 4945 generic.go:334] "Generic (PLEG): container finished" podID="63107fb2-24b7-4722-9e51-eae2ded28727" containerID="349e7b88ab04ff0e4e91f2816c8337d136ce652f40ae1724245d42edae13ad15" exitCode=0 Dec 06 08:19:27 crc kubenswrapper[4945]: I1206 08:19:27.459538 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dd684579-jss4d" event={"ID":"63107fb2-24b7-4722-9e51-eae2ded28727","Type":"ContainerDied","Data":"349e7b88ab04ff0e4e91f2816c8337d136ce652f40ae1724245d42edae13ad15"} Dec 06 08:19:27 crc kubenswrapper[4945]: I1206 08:19:27.484421 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d89777897-g2pzk" podStartSLOduration=2.484396927 podStartE2EDuration="2.484396927s" podCreationTimestamp="2025-12-06 08:19:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:19:27.478098638 +0000 UTC m=+5180.932959682" watchObservedRunningTime="2025-12-06 08:19:27.484396927 +0000 UTC m=+5180.939257971" Dec 06 08:19:28 crc kubenswrapper[4945]: I1206 08:19:28.469677 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dd684579-jss4d" event={"ID":"63107fb2-24b7-4722-9e51-eae2ded28727","Type":"ContainerStarted","Data":"65c0fe232277b6f2bff9669989dcb852fd220f497db671900a3ad29af64410b8"} Dec 06 08:19:28 crc kubenswrapper[4945]: I1206 08:19:28.469779 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:35 crc kubenswrapper[4945]: I1206 08:19:35.941415 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:19:35 crc kubenswrapper[4945]: I1206 08:19:35.962930 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64dd684579-jss4d" podStartSLOduration=10.962905228 podStartE2EDuration="10.962905228s" podCreationTimestamp="2025-12-06 08:19:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:19:28.488082218 +0000 UTC m=+5181.942943282" watchObservedRunningTime="2025-12-06 08:19:35.962905228 +0000 UTC m=+5189.417766302" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.012132 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6679b8d799-n8gmh"] Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.012466 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" podUID="f0f86a9a-4f9b-487a-98a6-6fb733642edb" containerName="dnsmasq-dns" containerID="cri-o://ababc9b1e1f5da53bf58818c8a38eb635e05c168eacca390ea6da51a37d60f2a" gracePeriod=10 Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.535710 4945 generic.go:334] "Generic (PLEG): container finished" podID="f0f86a9a-4f9b-487a-98a6-6fb733642edb" containerID="ababc9b1e1f5da53bf58818c8a38eb635e05c168eacca390ea6da51a37d60f2a" exitCode=0 Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.535807 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" event={"ID":"f0f86a9a-4f9b-487a-98a6-6fb733642edb","Type":"ContainerDied","Data":"ababc9b1e1f5da53bf58818c8a38eb635e05c168eacca390ea6da51a37d60f2a"} Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.536181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" event={"ID":"f0f86a9a-4f9b-487a-98a6-6fb733642edb","Type":"ContainerDied","Data":"83c97b9e4159a42f9b334f2e6747160b00b5c73346f315e094d64c0dad2a2550"} Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.536258 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83c97b9e4159a42f9b334f2e6747160b00b5c73346f315e094d64c0dad2a2550" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.546559 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.652757 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-dns-svc\") pod \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.652827 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-sb\") pod \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.652962 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-config\") pod \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.653020 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-nb\") pod \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.653042 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cstx6\" (UniqueName: \"kubernetes.io/projected/f0f86a9a-4f9b-487a-98a6-6fb733642edb-kube-api-access-cstx6\") pod \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\" (UID: \"f0f86a9a-4f9b-487a-98a6-6fb733642edb\") " Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.682544 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0f86a9a-4f9b-487a-98a6-6fb733642edb-kube-api-access-cstx6" (OuterVolumeSpecName: "kube-api-access-cstx6") pod "f0f86a9a-4f9b-487a-98a6-6fb733642edb" (UID: "f0f86a9a-4f9b-487a-98a6-6fb733642edb"). InnerVolumeSpecName "kube-api-access-cstx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.734475 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-config" (OuterVolumeSpecName: "config") pod "f0f86a9a-4f9b-487a-98a6-6fb733642edb" (UID: "f0f86a9a-4f9b-487a-98a6-6fb733642edb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.734919 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f0f86a9a-4f9b-487a-98a6-6fb733642edb" (UID: "f0f86a9a-4f9b-487a-98a6-6fb733642edb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.742780 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f0f86a9a-4f9b-487a-98a6-6fb733642edb" (UID: "f0f86a9a-4f9b-487a-98a6-6fb733642edb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.752554 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f0f86a9a-4f9b-487a-98a6-6fb733642edb" (UID: "f0f86a9a-4f9b-487a-98a6-6fb733642edb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.755472 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.755509 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cstx6\" (UniqueName: \"kubernetes.io/projected/f0f86a9a-4f9b-487a-98a6-6fb733642edb-kube-api-access-cstx6\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.755521 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.755530 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:36 crc kubenswrapper[4945]: I1206 08:19:36.755539 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0f86a9a-4f9b-487a-98a6-6fb733642edb-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:19:37 crc kubenswrapper[4945]: I1206 08:19:37.543789 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6679b8d799-n8gmh" Dec 06 08:19:37 crc kubenswrapper[4945]: I1206 08:19:37.566146 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6679b8d799-n8gmh"] Dec 06 08:19:37 crc kubenswrapper[4945]: I1206 08:19:37.575205 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6679b8d799-n8gmh"] Dec 06 08:19:38 crc kubenswrapper[4945]: I1206 08:19:38.963465 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0f86a9a-4f9b-487a-98a6-6fb733642edb" path="/var/lib/kubelet/pods/f0f86a9a-4f9b-487a-98a6-6fb733642edb/volumes" Dec 06 08:19:56 crc kubenswrapper[4945]: I1206 08:19:56.140459 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d89777897-g2pzk" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.557557 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-8qccw"] Dec 06 08:20:03 crc kubenswrapper[4945]: E1206 08:20:03.558656 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0f86a9a-4f9b-487a-98a6-6fb733642edb" containerName="dnsmasq-dns" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.558678 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0f86a9a-4f9b-487a-98a6-6fb733642edb" containerName="dnsmasq-dns" Dec 06 08:20:03 crc kubenswrapper[4945]: E1206 08:20:03.558722 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0f86a9a-4f9b-487a-98a6-6fb733642edb" containerName="init" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.558732 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0f86a9a-4f9b-487a-98a6-6fb733642edb" containerName="init" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.558937 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0f86a9a-4f9b-487a-98a6-6fb733642edb" containerName="dnsmasq-dns" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.559885 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8qccw" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.570298 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8qccw"] Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.658831 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a32bacdd-7760-40d0-ae80-b16857e2c03a-operator-scripts\") pod \"glance-db-create-8qccw\" (UID: \"a32bacdd-7760-40d0-ae80-b16857e2c03a\") " pod="openstack/glance-db-create-8qccw" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.658989 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6mkt\" (UniqueName: \"kubernetes.io/projected/a32bacdd-7760-40d0-ae80-b16857e2c03a-kube-api-access-w6mkt\") pod \"glance-db-create-8qccw\" (UID: \"a32bacdd-7760-40d0-ae80-b16857e2c03a\") " pod="openstack/glance-db-create-8qccw" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.674515 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b3a2-account-create-update-mn7tt"] Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.676972 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b3a2-account-create-update-mn7tt" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.683713 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.692615 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b3a2-account-create-update-mn7tt"] Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.761035 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a32bacdd-7760-40d0-ae80-b16857e2c03a-operator-scripts\") pod \"glance-db-create-8qccw\" (UID: \"a32bacdd-7760-40d0-ae80-b16857e2c03a\") " pod="openstack/glance-db-create-8qccw" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.761111 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls5hw\" (UniqueName: \"kubernetes.io/projected/3c788c88-bfb8-4ce3-8316-3ad663f44602-kube-api-access-ls5hw\") pod \"glance-b3a2-account-create-update-mn7tt\" (UID: \"3c788c88-bfb8-4ce3-8316-3ad663f44602\") " pod="openstack/glance-b3a2-account-create-update-mn7tt" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.761153 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c788c88-bfb8-4ce3-8316-3ad663f44602-operator-scripts\") pod \"glance-b3a2-account-create-update-mn7tt\" (UID: \"3c788c88-bfb8-4ce3-8316-3ad663f44602\") " pod="openstack/glance-b3a2-account-create-update-mn7tt" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.761203 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6mkt\" (UniqueName: \"kubernetes.io/projected/a32bacdd-7760-40d0-ae80-b16857e2c03a-kube-api-access-w6mkt\") pod \"glance-db-create-8qccw\" (UID: \"a32bacdd-7760-40d0-ae80-b16857e2c03a\") " pod="openstack/glance-db-create-8qccw" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.761849 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a32bacdd-7760-40d0-ae80-b16857e2c03a-operator-scripts\") pod \"glance-db-create-8qccw\" (UID: \"a32bacdd-7760-40d0-ae80-b16857e2c03a\") " pod="openstack/glance-db-create-8qccw" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.780105 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6mkt\" (UniqueName: \"kubernetes.io/projected/a32bacdd-7760-40d0-ae80-b16857e2c03a-kube-api-access-w6mkt\") pod \"glance-db-create-8qccw\" (UID: \"a32bacdd-7760-40d0-ae80-b16857e2c03a\") " pod="openstack/glance-db-create-8qccw" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.862205 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls5hw\" (UniqueName: \"kubernetes.io/projected/3c788c88-bfb8-4ce3-8316-3ad663f44602-kube-api-access-ls5hw\") pod \"glance-b3a2-account-create-update-mn7tt\" (UID: \"3c788c88-bfb8-4ce3-8316-3ad663f44602\") " pod="openstack/glance-b3a2-account-create-update-mn7tt" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.862291 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c788c88-bfb8-4ce3-8316-3ad663f44602-operator-scripts\") pod \"glance-b3a2-account-create-update-mn7tt\" (UID: \"3c788c88-bfb8-4ce3-8316-3ad663f44602\") " pod="openstack/glance-b3a2-account-create-update-mn7tt" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.863248 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c788c88-bfb8-4ce3-8316-3ad663f44602-operator-scripts\") pod \"glance-b3a2-account-create-update-mn7tt\" (UID: \"3c788c88-bfb8-4ce3-8316-3ad663f44602\") " pod="openstack/glance-b3a2-account-create-update-mn7tt" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.877959 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls5hw\" (UniqueName: \"kubernetes.io/projected/3c788c88-bfb8-4ce3-8316-3ad663f44602-kube-api-access-ls5hw\") pod \"glance-b3a2-account-create-update-mn7tt\" (UID: \"3c788c88-bfb8-4ce3-8316-3ad663f44602\") " pod="openstack/glance-b3a2-account-create-update-mn7tt" Dec 06 08:20:03 crc kubenswrapper[4945]: I1206 08:20:03.886122 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8qccw" Dec 06 08:20:04 crc kubenswrapper[4945]: I1206 08:20:04.003253 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b3a2-account-create-update-mn7tt" Dec 06 08:20:04 crc kubenswrapper[4945]: I1206 08:20:04.360069 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8qccw"] Dec 06 08:20:04 crc kubenswrapper[4945]: I1206 08:20:04.460267 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b3a2-account-create-update-mn7tt"] Dec 06 08:20:04 crc kubenswrapper[4945]: W1206 08:20:04.464070 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c788c88_bfb8_4ce3_8316_3ad663f44602.slice/crio-3927ca08903ffbde208a66ac2e44b7369c7f8e021e691544fa5f5ea043dffae7 WatchSource:0}: Error finding container 3927ca08903ffbde208a66ac2e44b7369c7f8e021e691544fa5f5ea043dffae7: Status 404 returned error can't find the container with id 3927ca08903ffbde208a66ac2e44b7369c7f8e021e691544fa5f5ea043dffae7 Dec 06 08:20:04 crc kubenswrapper[4945]: I1206 08:20:04.766132 4945 generic.go:334] "Generic (PLEG): container finished" podID="a32bacdd-7760-40d0-ae80-b16857e2c03a" containerID="d980bc39cee707bd1d99b68e271d88cf072743d74d782f674b630186e1297e0c" exitCode=0 Dec 06 08:20:04 crc kubenswrapper[4945]: I1206 08:20:04.766549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8qccw" event={"ID":"a32bacdd-7760-40d0-ae80-b16857e2c03a","Type":"ContainerDied","Data":"d980bc39cee707bd1d99b68e271d88cf072743d74d782f674b630186e1297e0c"} Dec 06 08:20:04 crc kubenswrapper[4945]: I1206 08:20:04.766576 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8qccw" event={"ID":"a32bacdd-7760-40d0-ae80-b16857e2c03a","Type":"ContainerStarted","Data":"ff26e9ab15c18ed564885f3ca817ab555552f8528efd2d90633be68016a224db"} Dec 06 08:20:04 crc kubenswrapper[4945]: I1206 08:20:04.768644 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b3a2-account-create-update-mn7tt" event={"ID":"3c788c88-bfb8-4ce3-8316-3ad663f44602","Type":"ContainerStarted","Data":"5b8dcd82f9b3b9ca89221d42f990e087e941cef295251973dd53c05c2aec8aa7"} Dec 06 08:20:04 crc kubenswrapper[4945]: I1206 08:20:04.768670 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b3a2-account-create-update-mn7tt" event={"ID":"3c788c88-bfb8-4ce3-8316-3ad663f44602","Type":"ContainerStarted","Data":"3927ca08903ffbde208a66ac2e44b7369c7f8e021e691544fa5f5ea043dffae7"} Dec 06 08:20:05 crc kubenswrapper[4945]: I1206 08:20:05.777750 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c788c88-bfb8-4ce3-8316-3ad663f44602" containerID="5b8dcd82f9b3b9ca89221d42f990e087e941cef295251973dd53c05c2aec8aa7" exitCode=0 Dec 06 08:20:05 crc kubenswrapper[4945]: I1206 08:20:05.777824 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b3a2-account-create-update-mn7tt" event={"ID":"3c788c88-bfb8-4ce3-8316-3ad663f44602","Type":"ContainerDied","Data":"5b8dcd82f9b3b9ca89221d42f990e087e941cef295251973dd53c05c2aec8aa7"} Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.169623 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8qccw" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.178453 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b3a2-account-create-update-mn7tt" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.304909 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6mkt\" (UniqueName: \"kubernetes.io/projected/a32bacdd-7760-40d0-ae80-b16857e2c03a-kube-api-access-w6mkt\") pod \"a32bacdd-7760-40d0-ae80-b16857e2c03a\" (UID: \"a32bacdd-7760-40d0-ae80-b16857e2c03a\") " Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.305096 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a32bacdd-7760-40d0-ae80-b16857e2c03a-operator-scripts\") pod \"a32bacdd-7760-40d0-ae80-b16857e2c03a\" (UID: \"a32bacdd-7760-40d0-ae80-b16857e2c03a\") " Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.305143 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c788c88-bfb8-4ce3-8316-3ad663f44602-operator-scripts\") pod \"3c788c88-bfb8-4ce3-8316-3ad663f44602\" (UID: \"3c788c88-bfb8-4ce3-8316-3ad663f44602\") " Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.305193 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls5hw\" (UniqueName: \"kubernetes.io/projected/3c788c88-bfb8-4ce3-8316-3ad663f44602-kube-api-access-ls5hw\") pod \"3c788c88-bfb8-4ce3-8316-3ad663f44602\" (UID: \"3c788c88-bfb8-4ce3-8316-3ad663f44602\") " Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.305793 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c788c88-bfb8-4ce3-8316-3ad663f44602-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c788c88-bfb8-4ce3-8316-3ad663f44602" (UID: "3c788c88-bfb8-4ce3-8316-3ad663f44602"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.306078 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a32bacdd-7760-40d0-ae80-b16857e2c03a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a32bacdd-7760-40d0-ae80-b16857e2c03a" (UID: "a32bacdd-7760-40d0-ae80-b16857e2c03a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.311043 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a32bacdd-7760-40d0-ae80-b16857e2c03a-kube-api-access-w6mkt" (OuterVolumeSpecName: "kube-api-access-w6mkt") pod "a32bacdd-7760-40d0-ae80-b16857e2c03a" (UID: "a32bacdd-7760-40d0-ae80-b16857e2c03a"). InnerVolumeSpecName "kube-api-access-w6mkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.311361 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c788c88-bfb8-4ce3-8316-3ad663f44602-kube-api-access-ls5hw" (OuterVolumeSpecName: "kube-api-access-ls5hw") pod "3c788c88-bfb8-4ce3-8316-3ad663f44602" (UID: "3c788c88-bfb8-4ce3-8316-3ad663f44602"). InnerVolumeSpecName "kube-api-access-ls5hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.407595 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a32bacdd-7760-40d0-ae80-b16857e2c03a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.407657 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c788c88-bfb8-4ce3-8316-3ad663f44602-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.407668 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls5hw\" (UniqueName: \"kubernetes.io/projected/3c788c88-bfb8-4ce3-8316-3ad663f44602-kube-api-access-ls5hw\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.407679 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6mkt\" (UniqueName: \"kubernetes.io/projected/a32bacdd-7760-40d0-ae80-b16857e2c03a-kube-api-access-w6mkt\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.787582 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b3a2-account-create-update-mn7tt" event={"ID":"3c788c88-bfb8-4ce3-8316-3ad663f44602","Type":"ContainerDied","Data":"3927ca08903ffbde208a66ac2e44b7369c7f8e021e691544fa5f5ea043dffae7"} Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.787652 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3927ca08903ffbde208a66ac2e44b7369c7f8e021e691544fa5f5ea043dffae7" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.787716 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b3a2-account-create-update-mn7tt" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.791627 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8qccw" event={"ID":"a32bacdd-7760-40d0-ae80-b16857e2c03a","Type":"ContainerDied","Data":"ff26e9ab15c18ed564885f3ca817ab555552f8528efd2d90633be68016a224db"} Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.791682 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8qccw" Dec 06 08:20:06 crc kubenswrapper[4945]: I1206 08:20:06.791703 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff26e9ab15c18ed564885f3ca817ab555552f8528efd2d90633be68016a224db" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.796122 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.796549 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.888518 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-nnq2k"] Dec 06 08:20:08 crc kubenswrapper[4945]: E1206 08:20:08.889107 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c788c88-bfb8-4ce3-8316-3ad663f44602" containerName="mariadb-account-create-update" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.889130 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c788c88-bfb8-4ce3-8316-3ad663f44602" containerName="mariadb-account-create-update" Dec 06 08:20:08 crc kubenswrapper[4945]: E1206 08:20:08.889157 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32bacdd-7760-40d0-ae80-b16857e2c03a" containerName="mariadb-database-create" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.889165 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32bacdd-7760-40d0-ae80-b16857e2c03a" containerName="mariadb-database-create" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.889396 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c788c88-bfb8-4ce3-8316-3ad663f44602" containerName="mariadb-account-create-update" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.889439 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32bacdd-7760-40d0-ae80-b16857e2c03a" containerName="mariadb-database-create" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.890302 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.892370 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-x8ldp" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.896683 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-nnq2k"] Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.897931 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.973084 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-combined-ca-bundle\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.973238 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcpv5\" (UniqueName: \"kubernetes.io/projected/7eb4c4a7-7e65-444c-b65b-206632d5037b-kube-api-access-dcpv5\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.973313 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-db-sync-config-data\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:08 crc kubenswrapper[4945]: I1206 08:20:08.973344 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-config-data\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.076031 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcpv5\" (UniqueName: \"kubernetes.io/projected/7eb4c4a7-7e65-444c-b65b-206632d5037b-kube-api-access-dcpv5\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.076138 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-db-sync-config-data\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.076165 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-config-data\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.076200 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-combined-ca-bundle\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.081885 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-db-sync-config-data\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.082056 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-config-data\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.091262 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcpv5\" (UniqueName: \"kubernetes.io/projected/7eb4c4a7-7e65-444c-b65b-206632d5037b-kube-api-access-dcpv5\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.095120 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-combined-ca-bundle\") pod \"glance-db-sync-nnq2k\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.229439 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.763130 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-nnq2k"] Dec 06 08:20:09 crc kubenswrapper[4945]: W1206 08:20:09.765106 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7eb4c4a7_7e65_444c_b65b_206632d5037b.slice/crio-afec742c40d20779898e2b65a1c9bef33527bfe150f350218f9d0cc760d0d3bc WatchSource:0}: Error finding container afec742c40d20779898e2b65a1c9bef33527bfe150f350218f9d0cc760d0d3bc: Status 404 returned error can't find the container with id afec742c40d20779898e2b65a1c9bef33527bfe150f350218f9d0cc760d0d3bc Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.767765 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 08:20:09 crc kubenswrapper[4945]: I1206 08:20:09.814560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nnq2k" event={"ID":"7eb4c4a7-7e65-444c-b65b-206632d5037b","Type":"ContainerStarted","Data":"afec742c40d20779898e2b65a1c9bef33527bfe150f350218f9d0cc760d0d3bc"} Dec 06 08:20:26 crc kubenswrapper[4945]: I1206 08:20:26.971539 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nnq2k" event={"ID":"7eb4c4a7-7e65-444c-b65b-206632d5037b","Type":"ContainerStarted","Data":"943434f63afced1af6932d562efb514a4cadbba4a237a0e7424949164d1e9fb9"} Dec 06 08:20:26 crc kubenswrapper[4945]: I1206 08:20:26.993626 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-nnq2k" podStartSLOduration=3.191490468 podStartE2EDuration="18.993605372s" podCreationTimestamp="2025-12-06 08:20:08 +0000 UTC" firstStartedPulling="2025-12-06 08:20:09.767436568 +0000 UTC m=+5223.222297622" lastFinishedPulling="2025-12-06 08:20:25.569551482 +0000 UTC m=+5239.024412526" observedRunningTime="2025-12-06 08:20:26.985696101 +0000 UTC m=+5240.440557145" watchObservedRunningTime="2025-12-06 08:20:26.993605372 +0000 UTC m=+5240.448466426" Dec 06 08:20:29 crc kubenswrapper[4945]: I1206 08:20:29.996684 4945 generic.go:334] "Generic (PLEG): container finished" podID="7eb4c4a7-7e65-444c-b65b-206632d5037b" containerID="943434f63afced1af6932d562efb514a4cadbba4a237a0e7424949164d1e9fb9" exitCode=0 Dec 06 08:20:29 crc kubenswrapper[4945]: I1206 08:20:29.996752 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nnq2k" event={"ID":"7eb4c4a7-7e65-444c-b65b-206632d5037b","Type":"ContainerDied","Data":"943434f63afced1af6932d562efb514a4cadbba4a237a0e7424949164d1e9fb9"} Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.364306 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.422960 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcpv5\" (UniqueName: \"kubernetes.io/projected/7eb4c4a7-7e65-444c-b65b-206632d5037b-kube-api-access-dcpv5\") pod \"7eb4c4a7-7e65-444c-b65b-206632d5037b\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.423117 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-db-sync-config-data\") pod \"7eb4c4a7-7e65-444c-b65b-206632d5037b\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.423185 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-config-data\") pod \"7eb4c4a7-7e65-444c-b65b-206632d5037b\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.423243 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-combined-ca-bundle\") pod \"7eb4c4a7-7e65-444c-b65b-206632d5037b\" (UID: \"7eb4c4a7-7e65-444c-b65b-206632d5037b\") " Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.429439 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7eb4c4a7-7e65-444c-b65b-206632d5037b" (UID: "7eb4c4a7-7e65-444c-b65b-206632d5037b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.429464 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb4c4a7-7e65-444c-b65b-206632d5037b-kube-api-access-dcpv5" (OuterVolumeSpecName: "kube-api-access-dcpv5") pod "7eb4c4a7-7e65-444c-b65b-206632d5037b" (UID: "7eb4c4a7-7e65-444c-b65b-206632d5037b"). InnerVolumeSpecName "kube-api-access-dcpv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.446641 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7eb4c4a7-7e65-444c-b65b-206632d5037b" (UID: "7eb4c4a7-7e65-444c-b65b-206632d5037b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.463958 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-config-data" (OuterVolumeSpecName: "config-data") pod "7eb4c4a7-7e65-444c-b65b-206632d5037b" (UID: "7eb4c4a7-7e65-444c-b65b-206632d5037b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.525463 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcpv5\" (UniqueName: \"kubernetes.io/projected/7eb4c4a7-7e65-444c-b65b-206632d5037b-kube-api-access-dcpv5\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.525497 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.525507 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:31 crc kubenswrapper[4945]: I1206 08:20:31.525517 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb4c4a7-7e65-444c-b65b-206632d5037b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.013077 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nnq2k" event={"ID":"7eb4c4a7-7e65-444c-b65b-206632d5037b","Type":"ContainerDied","Data":"afec742c40d20779898e2b65a1c9bef33527bfe150f350218f9d0cc760d0d3bc"} Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.013120 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afec742c40d20779898e2b65a1c9bef33527bfe150f350218f9d0cc760d0d3bc" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.013143 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nnq2k" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.377294 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:20:32 crc kubenswrapper[4945]: E1206 08:20:32.377968 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb4c4a7-7e65-444c-b65b-206632d5037b" containerName="glance-db-sync" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.377983 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb4c4a7-7e65-444c-b65b-206632d5037b" containerName="glance-db-sync" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.378156 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eb4c4a7-7e65-444c-b65b-206632d5037b" containerName="glance-db-sync" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.379106 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.398921 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.400981 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-x8ldp" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.400990 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.403625 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.406470 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.413472 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-759f69b595-gcn84"] Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.433180 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.441433 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-ceph\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.441548 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.441579 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.441602 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-logs\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.441677 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.441701 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl4g9\" (UniqueName: \"kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-kube-api-access-jl4g9\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.441752 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.454739 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759f69b595-gcn84"] Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.541603 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.543117 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544638 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-config\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544672 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-sb\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544690 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4p8b\" (UniqueName: \"kubernetes.io/projected/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-kube-api-access-l4p8b\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544733 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-ceph\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544757 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-nb\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544785 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544803 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544818 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-logs\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544865 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544880 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl4g9\" (UniqueName: \"kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-kube-api-access-jl4g9\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544915 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-dns-svc\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.544937 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.546190 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.547743 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-logs\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.547779 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.552195 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-ceph\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.552712 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.553533 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.554446 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.557132 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.581092 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl4g9\" (UniqueName: \"kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-kube-api-access-jl4g9\") pod \"glance-default-external-api-0\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.646111 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.646462 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bvlh\" (UniqueName: \"kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-kube-api-access-2bvlh\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.646633 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-nb\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.646876 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.647024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.647144 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.647252 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.647399 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-dns-svc\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.647504 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-logs\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.647622 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4p8b\" (UniqueName: \"kubernetes.io/projected/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-kube-api-access-l4p8b\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.647710 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-config\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.647799 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-sb\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.648254 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-dns-svc\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.647659 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-nb\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.649010 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-sb\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.649268 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-config\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.666767 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4p8b\" (UniqueName: \"kubernetes.io/projected/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-kube-api-access-l4p8b\") pod \"dnsmasq-dns-759f69b595-gcn84\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.700992 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.749536 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.749616 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.749658 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.749694 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.749741 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-logs\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.749780 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.749828 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bvlh\" (UniqueName: \"kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-kube-api-access-2bvlh\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.751823 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.754178 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-logs\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.756681 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.756850 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.757724 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.761838 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.762343 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.778819 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bvlh\" (UniqueName: \"kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-kube-api-access-2bvlh\") pod \"glance-default-internal-api-0\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:32 crc kubenswrapper[4945]: I1206 08:20:32.933657 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:33 crc kubenswrapper[4945]: I1206 08:20:33.320018 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:20:33 crc kubenswrapper[4945]: I1206 08:20:33.346133 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759f69b595-gcn84"] Dec 06 08:20:33 crc kubenswrapper[4945]: W1206 08:20:33.347068 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa27172f_213d_4a3a_b9df_6b6ad05de6b5.slice/crio-f091699a809c174f4827ebaec20733180794bee775392b05d698dfdde63af797 WatchSource:0}: Error finding container f091699a809c174f4827ebaec20733180794bee775392b05d698dfdde63af797: Status 404 returned error can't find the container with id f091699a809c174f4827ebaec20733180794bee775392b05d698dfdde63af797 Dec 06 08:20:33 crc kubenswrapper[4945]: I1206 08:20:33.508652 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:20:33 crc kubenswrapper[4945]: W1206 08:20:33.520790 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68d196fa_3d03_4402_9456_1400befd08e1.slice/crio-a67251c9c916c10c72e8ae31558bf63aea38dc11d63842eb053d3155e81c7830 WatchSource:0}: Error finding container a67251c9c916c10c72e8ae31558bf63aea38dc11d63842eb053d3155e81c7830: Status 404 returned error can't find the container with id a67251c9c916c10c72e8ae31558bf63aea38dc11d63842eb053d3155e81c7830 Dec 06 08:20:33 crc kubenswrapper[4945]: I1206 08:20:33.662000 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:20:34 crc kubenswrapper[4945]: I1206 08:20:34.034897 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"68d196fa-3d03-4402-9456-1400befd08e1","Type":"ContainerStarted","Data":"a67251c9c916c10c72e8ae31558bf63aea38dc11d63842eb053d3155e81c7830"} Dec 06 08:20:34 crc kubenswrapper[4945]: I1206 08:20:34.038478 4945 generic.go:334] "Generic (PLEG): container finished" podID="aa27172f-213d-4a3a-b9df-6b6ad05de6b5" containerID="0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336" exitCode=0 Dec 06 08:20:34 crc kubenswrapper[4945]: I1206 08:20:34.040070 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759f69b595-gcn84" event={"ID":"aa27172f-213d-4a3a-b9df-6b6ad05de6b5","Type":"ContainerDied","Data":"0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336"} Dec 06 08:20:34 crc kubenswrapper[4945]: I1206 08:20:34.040246 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759f69b595-gcn84" event={"ID":"aa27172f-213d-4a3a-b9df-6b6ad05de6b5","Type":"ContainerStarted","Data":"f091699a809c174f4827ebaec20733180794bee775392b05d698dfdde63af797"} Dec 06 08:20:34 crc kubenswrapper[4945]: I1206 08:20:34.045611 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373","Type":"ContainerStarted","Data":"7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc"} Dec 06 08:20:34 crc kubenswrapper[4945]: I1206 08:20:34.045652 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373","Type":"ContainerStarted","Data":"1ec4a7adf6525eb2928e744d5be4cfbd20887897c6a5d68e89f6adc7de188151"} Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.055174 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"68d196fa-3d03-4402-9456-1400befd08e1","Type":"ContainerStarted","Data":"8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147"} Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.055578 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"68d196fa-3d03-4402-9456-1400befd08e1","Type":"ContainerStarted","Data":"de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb"} Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.057390 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759f69b595-gcn84" event={"ID":"aa27172f-213d-4a3a-b9df-6b6ad05de6b5","Type":"ContainerStarted","Data":"baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7"} Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.057800 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.059537 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373","Type":"ContainerStarted","Data":"79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c"} Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.059633 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" containerName="glance-log" containerID="cri-o://7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc" gracePeriod=30 Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.059833 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" containerName="glance-httpd" containerID="cri-o://79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c" gracePeriod=30 Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.078618 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.078601483 podStartE2EDuration="3.078601483s" podCreationTimestamp="2025-12-06 08:20:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:20:35.074066342 +0000 UTC m=+5248.528927386" watchObservedRunningTime="2025-12-06 08:20:35.078601483 +0000 UTC m=+5248.533462527" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.129364 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.129346111 podStartE2EDuration="3.129346111s" podCreationTimestamp="2025-12-06 08:20:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:20:35.10201041 +0000 UTC m=+5248.556871454" watchObservedRunningTime="2025-12-06 08:20:35.129346111 +0000 UTC m=+5248.584207155" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.130357 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-759f69b595-gcn84" podStartSLOduration=3.130350788 podStartE2EDuration="3.130350788s" podCreationTimestamp="2025-12-06 08:20:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:20:35.125035126 +0000 UTC m=+5248.579896180" watchObservedRunningTime="2025-12-06 08:20:35.130350788 +0000 UTC m=+5248.585211832" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.704264 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.777420 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.815291 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-httpd-run\") pod \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.815380 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-logs\") pod \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.815429 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-config-data\") pod \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.815462 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-scripts\") pod \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.815486 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl4g9\" (UniqueName: \"kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-kube-api-access-jl4g9\") pod \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.815549 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-combined-ca-bundle\") pod \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.815591 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-ceph\") pod \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\" (UID: \"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373\") " Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.816026 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-logs" (OuterVolumeSpecName: "logs") pod "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" (UID: "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.816083 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" (UID: "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.823216 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-ceph" (OuterVolumeSpecName: "ceph") pod "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" (UID: "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.840727 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-scripts" (OuterVolumeSpecName: "scripts") pod "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" (UID: "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.846063 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" (UID: "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.847075 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-kube-api-access-jl4g9" (OuterVolumeSpecName: "kube-api-access-jl4g9") pod "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" (UID: "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373"). InnerVolumeSpecName "kube-api-access-jl4g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.888486 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-config-data" (OuterVolumeSpecName: "config-data") pod "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" (UID: "8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.917045 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.917218 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.917356 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl4g9\" (UniqueName: \"kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-kube-api-access-jl4g9\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.917447 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.917508 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.917597 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:35 crc kubenswrapper[4945]: I1206 08:20:35.917652 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.069338 4945 generic.go:334] "Generic (PLEG): container finished" podID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" containerID="79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c" exitCode=0 Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.069375 4945 generic.go:334] "Generic (PLEG): container finished" podID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" containerID="7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc" exitCode=143 Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.069423 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.069450 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373","Type":"ContainerDied","Data":"79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c"} Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.069495 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373","Type":"ContainerDied","Data":"7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc"} Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.069509 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373","Type":"ContainerDied","Data":"1ec4a7adf6525eb2928e744d5be4cfbd20887897c6a5d68e89f6adc7de188151"} Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.069524 4945 scope.go:117] "RemoveContainer" containerID="79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.094691 4945 scope.go:117] "RemoveContainer" containerID="7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.113985 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.130601 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.144615 4945 scope.go:117] "RemoveContainer" containerID="79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c" Dec 06 08:20:36 crc kubenswrapper[4945]: E1206 08:20:36.145481 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c\": container with ID starting with 79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c not found: ID does not exist" containerID="79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.145522 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c"} err="failed to get container status \"79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c\": rpc error: code = NotFound desc = could not find container \"79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c\": container with ID starting with 79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c not found: ID does not exist" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.145547 4945 scope.go:117] "RemoveContainer" containerID="7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc" Dec 06 08:20:36 crc kubenswrapper[4945]: E1206 08:20:36.152429 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc\": container with ID starting with 7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc not found: ID does not exist" containerID="7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.152476 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc"} err="failed to get container status \"7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc\": rpc error: code = NotFound desc = could not find container \"7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc\": container with ID starting with 7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc not found: ID does not exist" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.152507 4945 scope.go:117] "RemoveContainer" containerID="79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.156455 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c"} err="failed to get container status \"79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c\": rpc error: code = NotFound desc = could not find container \"79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c\": container with ID starting with 79ad308e4de3294b4efcff249a2d13db3c57edd73c8f50faf261d865a5f6eb8c not found: ID does not exist" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.156507 4945 scope.go:117] "RemoveContainer" containerID="7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.160475 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc"} err="failed to get container status \"7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc\": rpc error: code = NotFound desc = could not find container \"7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc\": container with ID starting with 7c4762e91785ff5f6bfa5fbec575c66c98369f315d5c66c719d8d87bfc63cfbc not found: ID does not exist" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.175597 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:20:36 crc kubenswrapper[4945]: E1206 08:20:36.176235 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" containerName="glance-log" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.176338 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" containerName="glance-log" Dec 06 08:20:36 crc kubenswrapper[4945]: E1206 08:20:36.176439 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" containerName="glance-httpd" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.176515 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" containerName="glance-httpd" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.176817 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" containerName="glance-log" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.176910 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" containerName="glance-httpd" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.178084 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.189319 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.222421 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.327247 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-ceph\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.327330 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-logs\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.327382 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.327450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-config-data\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.327474 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-scripts\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.327821 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx6kv\" (UniqueName: \"kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-kube-api-access-dx6kv\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.327845 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.429070 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-ceph\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.429127 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-logs\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.429167 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.429218 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-config-data\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.429236 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-scripts\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.429258 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx6kv\" (UniqueName: \"kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-kube-api-access-dx6kv\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.429277 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.429711 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-logs\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.429763 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.434199 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-ceph\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.434944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.435151 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-config-data\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.438973 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-scripts\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.449238 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx6kv\" (UniqueName: \"kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-kube-api-access-dx6kv\") pod \"glance-default-external-api-0\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.506896 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.892082 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:20:36 crc kubenswrapper[4945]: I1206 08:20:36.966803 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373" path="/var/lib/kubelet/pods/8a67e8cf-44f4-48a3-9eb7-ae0e5cb22373/volumes" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.082546 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c","Type":"ContainerStarted","Data":"35e98d7c50d2924403e245020eaff16deb69bef23bdf61abfd2f4d79bf289744"} Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.082781 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="68d196fa-3d03-4402-9456-1400befd08e1" containerName="glance-log" containerID="cri-o://de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb" gracePeriod=30 Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.082873 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="68d196fa-3d03-4402-9456-1400befd08e1" containerName="glance-httpd" containerID="cri-o://8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147" gracePeriod=30 Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.777587 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.858914 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-combined-ca-bundle\") pod \"68d196fa-3d03-4402-9456-1400befd08e1\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.859052 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-scripts\") pod \"68d196fa-3d03-4402-9456-1400befd08e1\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.859082 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-logs\") pod \"68d196fa-3d03-4402-9456-1400befd08e1\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.859108 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bvlh\" (UniqueName: \"kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-kube-api-access-2bvlh\") pod \"68d196fa-3d03-4402-9456-1400befd08e1\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.859136 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-config-data\") pod \"68d196fa-3d03-4402-9456-1400befd08e1\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.859156 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-httpd-run\") pod \"68d196fa-3d03-4402-9456-1400befd08e1\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.859171 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-ceph\") pod \"68d196fa-3d03-4402-9456-1400befd08e1\" (UID: \"68d196fa-3d03-4402-9456-1400befd08e1\") " Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.859898 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-logs" (OuterVolumeSpecName: "logs") pod "68d196fa-3d03-4402-9456-1400befd08e1" (UID: "68d196fa-3d03-4402-9456-1400befd08e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.860403 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "68d196fa-3d03-4402-9456-1400befd08e1" (UID: "68d196fa-3d03-4402-9456-1400befd08e1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.860791 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.860820 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/68d196fa-3d03-4402-9456-1400befd08e1-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.864141 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-scripts" (OuterVolumeSpecName: "scripts") pod "68d196fa-3d03-4402-9456-1400befd08e1" (UID: "68d196fa-3d03-4402-9456-1400befd08e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.865331 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-kube-api-access-2bvlh" (OuterVolumeSpecName: "kube-api-access-2bvlh") pod "68d196fa-3d03-4402-9456-1400befd08e1" (UID: "68d196fa-3d03-4402-9456-1400befd08e1"). InnerVolumeSpecName "kube-api-access-2bvlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.866022 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-ceph" (OuterVolumeSpecName: "ceph") pod "68d196fa-3d03-4402-9456-1400befd08e1" (UID: "68d196fa-3d03-4402-9456-1400befd08e1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.889180 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68d196fa-3d03-4402-9456-1400befd08e1" (UID: "68d196fa-3d03-4402-9456-1400befd08e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.919124 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-config-data" (OuterVolumeSpecName: "config-data") pod "68d196fa-3d03-4402-9456-1400befd08e1" (UID: "68d196fa-3d03-4402-9456-1400befd08e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.963603 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.963965 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.964042 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.964112 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d196fa-3d03-4402-9456-1400befd08e1-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:37 crc kubenswrapper[4945]: I1206 08:20:37.964177 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bvlh\" (UniqueName: \"kubernetes.io/projected/68d196fa-3d03-4402-9456-1400befd08e1-kube-api-access-2bvlh\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.093878 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c","Type":"ContainerStarted","Data":"2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42"} Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.098300 4945 generic.go:334] "Generic (PLEG): container finished" podID="68d196fa-3d03-4402-9456-1400befd08e1" containerID="8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147" exitCode=0 Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.098334 4945 generic.go:334] "Generic (PLEG): container finished" podID="68d196fa-3d03-4402-9456-1400befd08e1" containerID="de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb" exitCode=143 Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.098354 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"68d196fa-3d03-4402-9456-1400befd08e1","Type":"ContainerDied","Data":"8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147"} Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.098360 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.098375 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"68d196fa-3d03-4402-9456-1400befd08e1","Type":"ContainerDied","Data":"de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb"} Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.098385 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"68d196fa-3d03-4402-9456-1400befd08e1","Type":"ContainerDied","Data":"a67251c9c916c10c72e8ae31558bf63aea38dc11d63842eb053d3155e81c7830"} Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.098401 4945 scope.go:117] "RemoveContainer" containerID="8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.121228 4945 scope.go:117] "RemoveContainer" containerID="de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.138241 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.150815 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.162368 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:20:38 crc kubenswrapper[4945]: E1206 08:20:38.162861 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d196fa-3d03-4402-9456-1400befd08e1" containerName="glance-httpd" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.162878 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d196fa-3d03-4402-9456-1400befd08e1" containerName="glance-httpd" Dec 06 08:20:38 crc kubenswrapper[4945]: E1206 08:20:38.162913 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d196fa-3d03-4402-9456-1400befd08e1" containerName="glance-log" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.162920 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d196fa-3d03-4402-9456-1400befd08e1" containerName="glance-log" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.163117 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d196fa-3d03-4402-9456-1400befd08e1" containerName="glance-log" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.163144 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d196fa-3d03-4402-9456-1400befd08e1" containerName="glance-httpd" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.164267 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.168105 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.191031 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.195831 4945 scope.go:117] "RemoveContainer" containerID="8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147" Dec 06 08:20:38 crc kubenswrapper[4945]: E1206 08:20:38.197648 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147\": container with ID starting with 8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147 not found: ID does not exist" containerID="8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.197748 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147"} err="failed to get container status \"8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147\": rpc error: code = NotFound desc = could not find container \"8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147\": container with ID starting with 8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147 not found: ID does not exist" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.197827 4945 scope.go:117] "RemoveContainer" containerID="de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb" Dec 06 08:20:38 crc kubenswrapper[4945]: E1206 08:20:38.199391 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb\": container with ID starting with de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb not found: ID does not exist" containerID="de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.199487 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb"} err="failed to get container status \"de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb\": rpc error: code = NotFound desc = could not find container \"de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb\": container with ID starting with de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb not found: ID does not exist" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.199575 4945 scope.go:117] "RemoveContainer" containerID="8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.200527 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147"} err="failed to get container status \"8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147\": rpc error: code = NotFound desc = could not find container \"8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147\": container with ID starting with 8d70c6b137410b1b623ea1f184a2341a578ef967d406aca923c4880ddf9d5147 not found: ID does not exist" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.200614 4945 scope.go:117] "RemoveContainer" containerID="de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.202244 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb"} err="failed to get container status \"de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb\": rpc error: code = NotFound desc = could not find container \"de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb\": container with ID starting with de2a8ed9ed3ef9dd43bba1506c19e3e79ea4795e9422408ed6b735f0e6c4eeeb not found: ID does not exist" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.270082 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.270143 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.270161 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.270210 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-logs\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.270346 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.270765 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rphn\" (UniqueName: \"kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-kube-api-access-8rphn\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.270835 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.372268 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rphn\" (UniqueName: \"kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-kube-api-access-8rphn\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.372395 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.372419 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.372452 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.372468 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.373187 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-logs\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.373226 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.373557 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-logs\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.373762 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.377679 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.378781 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.379554 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.386078 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.398839 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rphn\" (UniqueName: \"kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-kube-api-access-8rphn\") pod \"glance-default-internal-api-0\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.489612 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.795840 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.796186 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:20:38 crc kubenswrapper[4945]: I1206 08:20:38.967397 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68d196fa-3d03-4402-9456-1400befd08e1" path="/var/lib/kubelet/pods/68d196fa-3d03-4402-9456-1400befd08e1/volumes" Dec 06 08:20:39 crc kubenswrapper[4945]: I1206 08:20:39.108635 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c","Type":"ContainerStarted","Data":"96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060"} Dec 06 08:20:39 crc kubenswrapper[4945]: I1206 08:20:39.130422 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.130407078 podStartE2EDuration="3.130407078s" podCreationTimestamp="2025-12-06 08:20:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:20:39.129187446 +0000 UTC m=+5252.584048490" watchObservedRunningTime="2025-12-06 08:20:39.130407078 +0000 UTC m=+5252.585268122" Dec 06 08:20:39 crc kubenswrapper[4945]: I1206 08:20:39.154948 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:20:40 crc kubenswrapper[4945]: I1206 08:20:40.119047 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6cb272d-4c37-4914-95e5-bcc1461b6ee9","Type":"ContainerStarted","Data":"750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a"} Dec 06 08:20:40 crc kubenswrapper[4945]: I1206 08:20:40.119855 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6cb272d-4c37-4914-95e5-bcc1461b6ee9","Type":"ContainerStarted","Data":"1bcff3830c59f5fef73ce7393f627bd603ae76a00c371d34e6a4561d9d5cbd87"} Dec 06 08:20:42 crc kubenswrapper[4945]: I1206 08:20:42.134956 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6cb272d-4c37-4914-95e5-bcc1461b6ee9","Type":"ContainerStarted","Data":"7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155"} Dec 06 08:20:42 crc kubenswrapper[4945]: I1206 08:20:42.160394 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.160362694 podStartE2EDuration="4.160362694s" podCreationTimestamp="2025-12-06 08:20:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:20:42.152994467 +0000 UTC m=+5255.607855531" watchObservedRunningTime="2025-12-06 08:20:42.160362694 +0000 UTC m=+5255.615223778" Dec 06 08:20:42 crc kubenswrapper[4945]: I1206 08:20:42.759311 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:20:42 crc kubenswrapper[4945]: I1206 08:20:42.813526 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64dd684579-jss4d"] Dec 06 08:20:42 crc kubenswrapper[4945]: I1206 08:20:42.813836 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64dd684579-jss4d" podUID="63107fb2-24b7-4722-9e51-eae2ded28727" containerName="dnsmasq-dns" containerID="cri-o://65c0fe232277b6f2bff9669989dcb852fd220f497db671900a3ad29af64410b8" gracePeriod=10 Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.175739 4945 generic.go:334] "Generic (PLEG): container finished" podID="63107fb2-24b7-4722-9e51-eae2ded28727" containerID="65c0fe232277b6f2bff9669989dcb852fd220f497db671900a3ad29af64410b8" exitCode=0 Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.176307 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dd684579-jss4d" event={"ID":"63107fb2-24b7-4722-9e51-eae2ded28727","Type":"ContainerDied","Data":"65c0fe232277b6f2bff9669989dcb852fd220f497db671900a3ad29af64410b8"} Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.379104 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.461326 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-config\") pod \"63107fb2-24b7-4722-9e51-eae2ded28727\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.461448 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29ptr\" (UniqueName: \"kubernetes.io/projected/63107fb2-24b7-4722-9e51-eae2ded28727-kube-api-access-29ptr\") pod \"63107fb2-24b7-4722-9e51-eae2ded28727\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.461497 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-sb\") pod \"63107fb2-24b7-4722-9e51-eae2ded28727\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.461517 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-nb\") pod \"63107fb2-24b7-4722-9e51-eae2ded28727\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.461577 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-dns-svc\") pod \"63107fb2-24b7-4722-9e51-eae2ded28727\" (UID: \"63107fb2-24b7-4722-9e51-eae2ded28727\") " Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.468418 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63107fb2-24b7-4722-9e51-eae2ded28727-kube-api-access-29ptr" (OuterVolumeSpecName: "kube-api-access-29ptr") pod "63107fb2-24b7-4722-9e51-eae2ded28727" (UID: "63107fb2-24b7-4722-9e51-eae2ded28727"). InnerVolumeSpecName "kube-api-access-29ptr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.502077 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "63107fb2-24b7-4722-9e51-eae2ded28727" (UID: "63107fb2-24b7-4722-9e51-eae2ded28727"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.502995 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "63107fb2-24b7-4722-9e51-eae2ded28727" (UID: "63107fb2-24b7-4722-9e51-eae2ded28727"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.503952 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-config" (OuterVolumeSpecName: "config") pod "63107fb2-24b7-4722-9e51-eae2ded28727" (UID: "63107fb2-24b7-4722-9e51-eae2ded28727"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.507862 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "63107fb2-24b7-4722-9e51-eae2ded28727" (UID: "63107fb2-24b7-4722-9e51-eae2ded28727"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.563896 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.563934 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29ptr\" (UniqueName: \"kubernetes.io/projected/63107fb2-24b7-4722-9e51-eae2ded28727-kube-api-access-29ptr\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.563948 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.563960 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:43 crc kubenswrapper[4945]: I1206 08:20:43.563968 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63107fb2-24b7-4722-9e51-eae2ded28727-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:20:44 crc kubenswrapper[4945]: I1206 08:20:44.187369 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dd684579-jss4d" event={"ID":"63107fb2-24b7-4722-9e51-eae2ded28727","Type":"ContainerDied","Data":"54b8d15a6d1df2116345ef5c898f2bd8df3ad028850a9b173f5af722c32948dc"} Dec 06 08:20:44 crc kubenswrapper[4945]: I1206 08:20:44.187442 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dd684579-jss4d" Dec 06 08:20:44 crc kubenswrapper[4945]: I1206 08:20:44.187456 4945 scope.go:117] "RemoveContainer" containerID="65c0fe232277b6f2bff9669989dcb852fd220f497db671900a3ad29af64410b8" Dec 06 08:20:44 crc kubenswrapper[4945]: I1206 08:20:44.228550 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64dd684579-jss4d"] Dec 06 08:20:44 crc kubenswrapper[4945]: I1206 08:20:44.228750 4945 scope.go:117] "RemoveContainer" containerID="349e7b88ab04ff0e4e91f2816c8337d136ce652f40ae1724245d42edae13ad15" Dec 06 08:20:44 crc kubenswrapper[4945]: I1206 08:20:44.239229 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64dd684579-jss4d"] Dec 06 08:20:44 crc kubenswrapper[4945]: E1206 08:20:44.281960 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63107fb2_24b7_4722_9e51_eae2ded28727.slice/crio-54b8d15a6d1df2116345ef5c898f2bd8df3ad028850a9b173f5af722c32948dc\": RecentStats: unable to find data in memory cache]" Dec 06 08:20:44 crc kubenswrapper[4945]: I1206 08:20:44.964676 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63107fb2-24b7-4722-9e51-eae2ded28727" path="/var/lib/kubelet/pods/63107fb2-24b7-4722-9e51-eae2ded28727/volumes" Dec 06 08:20:46 crc kubenswrapper[4945]: I1206 08:20:46.507464 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 08:20:46 crc kubenswrapper[4945]: I1206 08:20:46.507789 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 08:20:46 crc kubenswrapper[4945]: I1206 08:20:46.535948 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 08:20:46 crc kubenswrapper[4945]: I1206 08:20:46.570772 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 08:20:47 crc kubenswrapper[4945]: I1206 08:20:47.218664 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 08:20:47 crc kubenswrapper[4945]: I1206 08:20:47.219077 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 08:20:48 crc kubenswrapper[4945]: I1206 08:20:48.489881 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:48 crc kubenswrapper[4945]: I1206 08:20:48.490159 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:48 crc kubenswrapper[4945]: I1206 08:20:48.517558 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:48 crc kubenswrapper[4945]: I1206 08:20:48.534559 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:49 crc kubenswrapper[4945]: I1206 08:20:49.153985 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 08:20:49 crc kubenswrapper[4945]: I1206 08:20:49.157041 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 08:20:49 crc kubenswrapper[4945]: I1206 08:20:49.243652 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:49 crc kubenswrapper[4945]: I1206 08:20:49.243962 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:51 crc kubenswrapper[4945]: I1206 08:20:51.271108 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:51 crc kubenswrapper[4945]: I1206 08:20:51.271738 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 08:20:51 crc kubenswrapper[4945]: I1206 08:20:51.305555 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.143169 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-vr5x5"] Dec 06 08:20:59 crc kubenswrapper[4945]: E1206 08:20:59.144159 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63107fb2-24b7-4722-9e51-eae2ded28727" containerName="init" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.144173 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="63107fb2-24b7-4722-9e51-eae2ded28727" containerName="init" Dec 06 08:20:59 crc kubenswrapper[4945]: E1206 08:20:59.144195 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63107fb2-24b7-4722-9e51-eae2ded28727" containerName="dnsmasq-dns" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.144201 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="63107fb2-24b7-4722-9e51-eae2ded28727" containerName="dnsmasq-dns" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.144455 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="63107fb2-24b7-4722-9e51-eae2ded28727" containerName="dnsmasq-dns" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.145030 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vr5x5" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.156312 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vr5x5"] Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.250813 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg7td\" (UniqueName: \"kubernetes.io/projected/c1a1ef05-412d-41bb-be03-a2f649984250-kube-api-access-hg7td\") pod \"placement-db-create-vr5x5\" (UID: \"c1a1ef05-412d-41bb-be03-a2f649984250\") " pod="openstack/placement-db-create-vr5x5" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.251168 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1a1ef05-412d-41bb-be03-a2f649984250-operator-scripts\") pod \"placement-db-create-vr5x5\" (UID: \"c1a1ef05-412d-41bb-be03-a2f649984250\") " pod="openstack/placement-db-create-vr5x5" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.252346 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-26dd-account-create-update-h4ps4"] Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.253791 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-26dd-account-create-update-h4ps4" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.255827 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.262604 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-26dd-account-create-update-h4ps4"] Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.352849 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg7td\" (UniqueName: \"kubernetes.io/projected/c1a1ef05-412d-41bb-be03-a2f649984250-kube-api-access-hg7td\") pod \"placement-db-create-vr5x5\" (UID: \"c1a1ef05-412d-41bb-be03-a2f649984250\") " pod="openstack/placement-db-create-vr5x5" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.352931 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkhg9\" (UniqueName: \"kubernetes.io/projected/909d1589-58d9-412e-9b0e-15171d896b7b-kube-api-access-vkhg9\") pod \"placement-26dd-account-create-update-h4ps4\" (UID: \"909d1589-58d9-412e-9b0e-15171d896b7b\") " pod="openstack/placement-26dd-account-create-update-h4ps4" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.353151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1a1ef05-412d-41bb-be03-a2f649984250-operator-scripts\") pod \"placement-db-create-vr5x5\" (UID: \"c1a1ef05-412d-41bb-be03-a2f649984250\") " pod="openstack/placement-db-create-vr5x5" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.353246 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/909d1589-58d9-412e-9b0e-15171d896b7b-operator-scripts\") pod \"placement-26dd-account-create-update-h4ps4\" (UID: \"909d1589-58d9-412e-9b0e-15171d896b7b\") " pod="openstack/placement-26dd-account-create-update-h4ps4" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.353974 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1a1ef05-412d-41bb-be03-a2f649984250-operator-scripts\") pod \"placement-db-create-vr5x5\" (UID: \"c1a1ef05-412d-41bb-be03-a2f649984250\") " pod="openstack/placement-db-create-vr5x5" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.374059 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg7td\" (UniqueName: \"kubernetes.io/projected/c1a1ef05-412d-41bb-be03-a2f649984250-kube-api-access-hg7td\") pod \"placement-db-create-vr5x5\" (UID: \"c1a1ef05-412d-41bb-be03-a2f649984250\") " pod="openstack/placement-db-create-vr5x5" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.454450 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/909d1589-58d9-412e-9b0e-15171d896b7b-operator-scripts\") pod \"placement-26dd-account-create-update-h4ps4\" (UID: \"909d1589-58d9-412e-9b0e-15171d896b7b\") " pod="openstack/placement-26dd-account-create-update-h4ps4" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.454611 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkhg9\" (UniqueName: \"kubernetes.io/projected/909d1589-58d9-412e-9b0e-15171d896b7b-kube-api-access-vkhg9\") pod \"placement-26dd-account-create-update-h4ps4\" (UID: \"909d1589-58d9-412e-9b0e-15171d896b7b\") " pod="openstack/placement-26dd-account-create-update-h4ps4" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.455631 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/909d1589-58d9-412e-9b0e-15171d896b7b-operator-scripts\") pod \"placement-26dd-account-create-update-h4ps4\" (UID: \"909d1589-58d9-412e-9b0e-15171d896b7b\") " pod="openstack/placement-26dd-account-create-update-h4ps4" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.472022 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkhg9\" (UniqueName: \"kubernetes.io/projected/909d1589-58d9-412e-9b0e-15171d896b7b-kube-api-access-vkhg9\") pod \"placement-26dd-account-create-update-h4ps4\" (UID: \"909d1589-58d9-412e-9b0e-15171d896b7b\") " pod="openstack/placement-26dd-account-create-update-h4ps4" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.518263 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vr5x5" Dec 06 08:20:59 crc kubenswrapper[4945]: I1206 08:20:59.570298 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-26dd-account-create-update-h4ps4" Dec 06 08:21:00 crc kubenswrapper[4945]: I1206 08:21:00.072820 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vr5x5"] Dec 06 08:21:00 crc kubenswrapper[4945]: W1206 08:21:00.073206 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1a1ef05_412d_41bb_be03_a2f649984250.slice/crio-5234d6252a7e04bf6a3234f66a99bc27fc4d44f2be79d222727969145e67cb89 WatchSource:0}: Error finding container 5234d6252a7e04bf6a3234f66a99bc27fc4d44f2be79d222727969145e67cb89: Status 404 returned error can't find the container with id 5234d6252a7e04bf6a3234f66a99bc27fc4d44f2be79d222727969145e67cb89 Dec 06 08:21:00 crc kubenswrapper[4945]: W1206 08:21:00.207245 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod909d1589_58d9_412e_9b0e_15171d896b7b.slice/crio-ee74a4f14d4f3bcd18d0a175f4f8040229158ed3da16310f6f66de1b3f56c702 WatchSource:0}: Error finding container ee74a4f14d4f3bcd18d0a175f4f8040229158ed3da16310f6f66de1b3f56c702: Status 404 returned error can't find the container with id ee74a4f14d4f3bcd18d0a175f4f8040229158ed3da16310f6f66de1b3f56c702 Dec 06 08:21:00 crc kubenswrapper[4945]: I1206 08:21:00.209970 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-26dd-account-create-update-h4ps4"] Dec 06 08:21:00 crc kubenswrapper[4945]: I1206 08:21:00.328152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-26dd-account-create-update-h4ps4" event={"ID":"909d1589-58d9-412e-9b0e-15171d896b7b","Type":"ContainerStarted","Data":"ee74a4f14d4f3bcd18d0a175f4f8040229158ed3da16310f6f66de1b3f56c702"} Dec 06 08:21:00 crc kubenswrapper[4945]: I1206 08:21:00.329968 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vr5x5" event={"ID":"c1a1ef05-412d-41bb-be03-a2f649984250","Type":"ContainerStarted","Data":"d8e46074d4c0283a0851990cddde3645febfde2fd6047ae34a343c449cd2ef1f"} Dec 06 08:21:00 crc kubenswrapper[4945]: I1206 08:21:00.329998 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vr5x5" event={"ID":"c1a1ef05-412d-41bb-be03-a2f649984250","Type":"ContainerStarted","Data":"5234d6252a7e04bf6a3234f66a99bc27fc4d44f2be79d222727969145e67cb89"} Dec 06 08:21:01 crc kubenswrapper[4945]: I1206 08:21:01.339634 4945 generic.go:334] "Generic (PLEG): container finished" podID="909d1589-58d9-412e-9b0e-15171d896b7b" containerID="3eb2f8ee314c0830aadbbffe9617798c35e65645322e58766718d24477a80b16" exitCode=0 Dec 06 08:21:01 crc kubenswrapper[4945]: I1206 08:21:01.339885 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-26dd-account-create-update-h4ps4" event={"ID":"909d1589-58d9-412e-9b0e-15171d896b7b","Type":"ContainerDied","Data":"3eb2f8ee314c0830aadbbffe9617798c35e65645322e58766718d24477a80b16"} Dec 06 08:21:01 crc kubenswrapper[4945]: I1206 08:21:01.342824 4945 generic.go:334] "Generic (PLEG): container finished" podID="c1a1ef05-412d-41bb-be03-a2f649984250" containerID="d8e46074d4c0283a0851990cddde3645febfde2fd6047ae34a343c449cd2ef1f" exitCode=0 Dec 06 08:21:01 crc kubenswrapper[4945]: I1206 08:21:01.342926 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vr5x5" event={"ID":"c1a1ef05-412d-41bb-be03-a2f649984250","Type":"ContainerDied","Data":"d8e46074d4c0283a0851990cddde3645febfde2fd6047ae34a343c449cd2ef1f"} Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.729371 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vr5x5" Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.738960 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-26dd-account-create-update-h4ps4" Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.817420 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/909d1589-58d9-412e-9b0e-15171d896b7b-operator-scripts\") pod \"909d1589-58d9-412e-9b0e-15171d896b7b\" (UID: \"909d1589-58d9-412e-9b0e-15171d896b7b\") " Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.817506 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1a1ef05-412d-41bb-be03-a2f649984250-operator-scripts\") pod \"c1a1ef05-412d-41bb-be03-a2f649984250\" (UID: \"c1a1ef05-412d-41bb-be03-a2f649984250\") " Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.817595 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg7td\" (UniqueName: \"kubernetes.io/projected/c1a1ef05-412d-41bb-be03-a2f649984250-kube-api-access-hg7td\") pod \"c1a1ef05-412d-41bb-be03-a2f649984250\" (UID: \"c1a1ef05-412d-41bb-be03-a2f649984250\") " Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.817633 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkhg9\" (UniqueName: \"kubernetes.io/projected/909d1589-58d9-412e-9b0e-15171d896b7b-kube-api-access-vkhg9\") pod \"909d1589-58d9-412e-9b0e-15171d896b7b\" (UID: \"909d1589-58d9-412e-9b0e-15171d896b7b\") " Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.818176 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/909d1589-58d9-412e-9b0e-15171d896b7b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "909d1589-58d9-412e-9b0e-15171d896b7b" (UID: "909d1589-58d9-412e-9b0e-15171d896b7b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.818189 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a1ef05-412d-41bb-be03-a2f649984250-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c1a1ef05-412d-41bb-be03-a2f649984250" (UID: "c1a1ef05-412d-41bb-be03-a2f649984250"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.823987 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/909d1589-58d9-412e-9b0e-15171d896b7b-kube-api-access-vkhg9" (OuterVolumeSpecName: "kube-api-access-vkhg9") pod "909d1589-58d9-412e-9b0e-15171d896b7b" (UID: "909d1589-58d9-412e-9b0e-15171d896b7b"). InnerVolumeSpecName "kube-api-access-vkhg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.825453 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1a1ef05-412d-41bb-be03-a2f649984250-kube-api-access-hg7td" (OuterVolumeSpecName: "kube-api-access-hg7td") pod "c1a1ef05-412d-41bb-be03-a2f649984250" (UID: "c1a1ef05-412d-41bb-be03-a2f649984250"). InnerVolumeSpecName "kube-api-access-hg7td". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.920315 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg7td\" (UniqueName: \"kubernetes.io/projected/c1a1ef05-412d-41bb-be03-a2f649984250-kube-api-access-hg7td\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.920398 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkhg9\" (UniqueName: \"kubernetes.io/projected/909d1589-58d9-412e-9b0e-15171d896b7b-kube-api-access-vkhg9\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.920416 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/909d1589-58d9-412e-9b0e-15171d896b7b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:02 crc kubenswrapper[4945]: I1206 08:21:02.920435 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1a1ef05-412d-41bb-be03-a2f649984250-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:03 crc kubenswrapper[4945]: I1206 08:21:03.360490 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vr5x5" event={"ID":"c1a1ef05-412d-41bb-be03-a2f649984250","Type":"ContainerDied","Data":"5234d6252a7e04bf6a3234f66a99bc27fc4d44f2be79d222727969145e67cb89"} Dec 06 08:21:03 crc kubenswrapper[4945]: I1206 08:21:03.360773 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5234d6252a7e04bf6a3234f66a99bc27fc4d44f2be79d222727969145e67cb89" Dec 06 08:21:03 crc kubenswrapper[4945]: I1206 08:21:03.360528 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vr5x5" Dec 06 08:21:03 crc kubenswrapper[4945]: I1206 08:21:03.362199 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-26dd-account-create-update-h4ps4" event={"ID":"909d1589-58d9-412e-9b0e-15171d896b7b","Type":"ContainerDied","Data":"ee74a4f14d4f3bcd18d0a175f4f8040229158ed3da16310f6f66de1b3f56c702"} Dec 06 08:21:03 crc kubenswrapper[4945]: I1206 08:21:03.362230 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee74a4f14d4f3bcd18d0a175f4f8040229158ed3da16310f6f66de1b3f56c702" Dec 06 08:21:03 crc kubenswrapper[4945]: I1206 08:21:03.362268 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-26dd-account-create-update-h4ps4" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.536546 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b5d65bdf9-jv9rg"] Dec 06 08:21:04 crc kubenswrapper[4945]: E1206 08:21:04.537256 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909d1589-58d9-412e-9b0e-15171d896b7b" containerName="mariadb-account-create-update" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.537272 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="909d1589-58d9-412e-9b0e-15171d896b7b" containerName="mariadb-account-create-update" Dec 06 08:21:04 crc kubenswrapper[4945]: E1206 08:21:04.537318 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a1ef05-412d-41bb-be03-a2f649984250" containerName="mariadb-database-create" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.537325 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a1ef05-412d-41bb-be03-a2f649984250" containerName="mariadb-database-create" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.537491 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="909d1589-58d9-412e-9b0e-15171d896b7b" containerName="mariadb-account-create-update" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.537505 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1a1ef05-412d-41bb-be03-a2f649984250" containerName="mariadb-database-create" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.540086 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.560658 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b5d65bdf9-jv9rg"] Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.617731 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-5tkvt"] Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.621272 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.627012 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.627453 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zzh2l" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.627671 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.629895 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5tkvt"] Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.669289 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-sb\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.669380 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-dns-svc\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.669502 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-nb\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.669611 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71bda92d-e6dd-4061-874f-4020f86a223e-logs\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.669646 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-scripts\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.669671 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-config-data\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.669806 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-combined-ca-bundle\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.669839 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lfwh\" (UniqueName: \"kubernetes.io/projected/71bda92d-e6dd-4061-874f-4020f86a223e-kube-api-access-4lfwh\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.669877 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc64z\" (UniqueName: \"kubernetes.io/projected/9d508820-8bdf-4efe-9adc-d40c565b4378-kube-api-access-gc64z\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.669958 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-config\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.771045 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-config\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.771140 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-sb\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.771173 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-dns-svc\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.771197 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-nb\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.771219 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71bda92d-e6dd-4061-874f-4020f86a223e-logs\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.771240 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-scripts\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.771263 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-config-data\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.771337 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-combined-ca-bundle\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.771357 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lfwh\" (UniqueName: \"kubernetes.io/projected/71bda92d-e6dd-4061-874f-4020f86a223e-kube-api-access-4lfwh\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.771381 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc64z\" (UniqueName: \"kubernetes.io/projected/9d508820-8bdf-4efe-9adc-d40c565b4378-kube-api-access-gc64z\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.772019 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-config\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.772106 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-sb\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.772248 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-dns-svc\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.772744 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-nb\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.772786 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71bda92d-e6dd-4061-874f-4020f86a223e-logs\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.782818 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-scripts\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.783450 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-config-data\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.786484 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-combined-ca-bundle\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.788817 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lfwh\" (UniqueName: \"kubernetes.io/projected/71bda92d-e6dd-4061-874f-4020f86a223e-kube-api-access-4lfwh\") pod \"placement-db-sync-5tkvt\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.790650 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc64z\" (UniqueName: \"kubernetes.io/projected/9d508820-8bdf-4efe-9adc-d40c565b4378-kube-api-access-gc64z\") pod \"dnsmasq-dns-b5d65bdf9-jv9rg\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.872932 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:04 crc kubenswrapper[4945]: I1206 08:21:04.951513 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:05 crc kubenswrapper[4945]: I1206 08:21:05.345493 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b5d65bdf9-jv9rg"] Dec 06 08:21:05 crc kubenswrapper[4945]: I1206 08:21:05.380945 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" event={"ID":"9d508820-8bdf-4efe-9adc-d40c565b4378","Type":"ContainerStarted","Data":"62b898cdb2444c4fb453b1c71e9f4d6628427725d4fac57f2fb84bbe7c62976a"} Dec 06 08:21:05 crc kubenswrapper[4945]: I1206 08:21:05.438857 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5tkvt"] Dec 06 08:21:05 crc kubenswrapper[4945]: W1206 08:21:05.453125 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71bda92d_e6dd_4061_874f_4020f86a223e.slice/crio-7be1ab3b9dbdff2c9f614db3763a9f1d8b6123d2c8a2ed812653591fc7a77ba8 WatchSource:0}: Error finding container 7be1ab3b9dbdff2c9f614db3763a9f1d8b6123d2c8a2ed812653591fc7a77ba8: Status 404 returned error can't find the container with id 7be1ab3b9dbdff2c9f614db3763a9f1d8b6123d2c8a2ed812653591fc7a77ba8 Dec 06 08:21:06 crc kubenswrapper[4945]: I1206 08:21:06.391097 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5tkvt" event={"ID":"71bda92d-e6dd-4061-874f-4020f86a223e","Type":"ContainerStarted","Data":"7be1ab3b9dbdff2c9f614db3763a9f1d8b6123d2c8a2ed812653591fc7a77ba8"} Dec 06 08:21:06 crc kubenswrapper[4945]: I1206 08:21:06.394319 4945 generic.go:334] "Generic (PLEG): container finished" podID="9d508820-8bdf-4efe-9adc-d40c565b4378" containerID="068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7" exitCode=0 Dec 06 08:21:06 crc kubenswrapper[4945]: I1206 08:21:06.394381 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" event={"ID":"9d508820-8bdf-4efe-9adc-d40c565b4378","Type":"ContainerDied","Data":"068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7"} Dec 06 08:21:07 crc kubenswrapper[4945]: I1206 08:21:07.405174 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" event={"ID":"9d508820-8bdf-4efe-9adc-d40c565b4378","Type":"ContainerStarted","Data":"bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5"} Dec 06 08:21:07 crc kubenswrapper[4945]: I1206 08:21:07.405944 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:07 crc kubenswrapper[4945]: I1206 08:21:07.423306 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" podStartSLOduration=3.423270227 podStartE2EDuration="3.423270227s" podCreationTimestamp="2025-12-06 08:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:21:07.420929274 +0000 UTC m=+5280.875790338" watchObservedRunningTime="2025-12-06 08:21:07.423270227 +0000 UTC m=+5280.878131271" Dec 06 08:21:08 crc kubenswrapper[4945]: I1206 08:21:08.796205 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:21:08 crc kubenswrapper[4945]: I1206 08:21:08.798140 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:21:08 crc kubenswrapper[4945]: I1206 08:21:08.798185 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 08:21:08 crc kubenswrapper[4945]: I1206 08:21:08.798855 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 08:21:08 crc kubenswrapper[4945]: I1206 08:21:08.798905 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" gracePeriod=600 Dec 06 08:21:09 crc kubenswrapper[4945]: E1206 08:21:09.817000 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:21:10 crc kubenswrapper[4945]: I1206 08:21:10.434075 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5tkvt" event={"ID":"71bda92d-e6dd-4061-874f-4020f86a223e","Type":"ContainerStarted","Data":"6c6a9a23d4d92392bb44d53ab56d1faff17e9683c0970184f7b530749924d9cf"} Dec 06 08:21:10 crc kubenswrapper[4945]: I1206 08:21:10.439264 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" exitCode=0 Dec 06 08:21:10 crc kubenswrapper[4945]: I1206 08:21:10.439342 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39"} Dec 06 08:21:10 crc kubenswrapper[4945]: I1206 08:21:10.439377 4945 scope.go:117] "RemoveContainer" containerID="8e73a93782f4167b2d20799f17eea711a9f4910590603bc3ce04323932093a55" Dec 06 08:21:10 crc kubenswrapper[4945]: I1206 08:21:10.439985 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:21:10 crc kubenswrapper[4945]: E1206 08:21:10.440259 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:21:10 crc kubenswrapper[4945]: I1206 08:21:10.465476 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-5tkvt" podStartSLOduration=2.049332207 podStartE2EDuration="6.465449619s" podCreationTimestamp="2025-12-06 08:21:04 +0000 UTC" firstStartedPulling="2025-12-06 08:21:05.456088036 +0000 UTC m=+5278.910949080" lastFinishedPulling="2025-12-06 08:21:09.872205448 +0000 UTC m=+5283.327066492" observedRunningTime="2025-12-06 08:21:10.45202675 +0000 UTC m=+5283.906887824" watchObservedRunningTime="2025-12-06 08:21:10.465449619 +0000 UTC m=+5283.920310693" Dec 06 08:21:11 crc kubenswrapper[4945]: I1206 08:21:11.449665 4945 generic.go:334] "Generic (PLEG): container finished" podID="71bda92d-e6dd-4061-874f-4020f86a223e" containerID="6c6a9a23d4d92392bb44d53ab56d1faff17e9683c0970184f7b530749924d9cf" exitCode=0 Dec 06 08:21:11 crc kubenswrapper[4945]: I1206 08:21:11.449727 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5tkvt" event={"ID":"71bda92d-e6dd-4061-874f-4020f86a223e","Type":"ContainerDied","Data":"6c6a9a23d4d92392bb44d53ab56d1faff17e9683c0970184f7b530749924d9cf"} Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.789405 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.920735 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-config-data\") pod \"71bda92d-e6dd-4061-874f-4020f86a223e\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.920785 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-combined-ca-bundle\") pod \"71bda92d-e6dd-4061-874f-4020f86a223e\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.920814 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-scripts\") pod \"71bda92d-e6dd-4061-874f-4020f86a223e\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.920894 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71bda92d-e6dd-4061-874f-4020f86a223e-logs\") pod \"71bda92d-e6dd-4061-874f-4020f86a223e\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.920946 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lfwh\" (UniqueName: \"kubernetes.io/projected/71bda92d-e6dd-4061-874f-4020f86a223e-kube-api-access-4lfwh\") pod \"71bda92d-e6dd-4061-874f-4020f86a223e\" (UID: \"71bda92d-e6dd-4061-874f-4020f86a223e\") " Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.922231 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71bda92d-e6dd-4061-874f-4020f86a223e-logs" (OuterVolumeSpecName: "logs") pod "71bda92d-e6dd-4061-874f-4020f86a223e" (UID: "71bda92d-e6dd-4061-874f-4020f86a223e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.926296 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-scripts" (OuterVolumeSpecName: "scripts") pod "71bda92d-e6dd-4061-874f-4020f86a223e" (UID: "71bda92d-e6dd-4061-874f-4020f86a223e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.936385 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71bda92d-e6dd-4061-874f-4020f86a223e-kube-api-access-4lfwh" (OuterVolumeSpecName: "kube-api-access-4lfwh") pod "71bda92d-e6dd-4061-874f-4020f86a223e" (UID: "71bda92d-e6dd-4061-874f-4020f86a223e"). InnerVolumeSpecName "kube-api-access-4lfwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.949314 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-config-data" (OuterVolumeSpecName: "config-data") pod "71bda92d-e6dd-4061-874f-4020f86a223e" (UID: "71bda92d-e6dd-4061-874f-4020f86a223e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:21:12 crc kubenswrapper[4945]: I1206 08:21:12.954591 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71bda92d-e6dd-4061-874f-4020f86a223e" (UID: "71bda92d-e6dd-4061-874f-4020f86a223e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.023855 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.023895 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.023907 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71bda92d-e6dd-4061-874f-4020f86a223e-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.023917 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71bda92d-e6dd-4061-874f-4020f86a223e-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.023926 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lfwh\" (UniqueName: \"kubernetes.io/projected/71bda92d-e6dd-4061-874f-4020f86a223e-kube-api-access-4lfwh\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.465067 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5tkvt" event={"ID":"71bda92d-e6dd-4061-874f-4020f86a223e","Type":"ContainerDied","Data":"7be1ab3b9dbdff2c9f614db3763a9f1d8b6123d2c8a2ed812653591fc7a77ba8"} Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.465106 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5tkvt" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.465116 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7be1ab3b9dbdff2c9f614db3763a9f1d8b6123d2c8a2ed812653591fc7a77ba8" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.541209 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5c8c7ff5f6-w4pdr"] Dec 06 08:21:13 crc kubenswrapper[4945]: E1206 08:21:13.541616 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71bda92d-e6dd-4061-874f-4020f86a223e" containerName="placement-db-sync" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.541633 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="71bda92d-e6dd-4061-874f-4020f86a223e" containerName="placement-db-sync" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.541856 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="71bda92d-e6dd-4061-874f-4020f86a223e" containerName="placement-db-sync" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.542782 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.545222 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.546131 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.546443 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zzh2l" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.560628 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5c8c7ff5f6-w4pdr"] Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.631099 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5cb978-dca7-4d06-8c84-c06588c6eacd-logs\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.631154 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5cb978-dca7-4d06-8c84-c06588c6eacd-combined-ca-bundle\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.631231 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5cb978-dca7-4d06-8c84-c06588c6eacd-scripts\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.631289 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2bs4\" (UniqueName: \"kubernetes.io/projected/8c5cb978-dca7-4d06-8c84-c06588c6eacd-kube-api-access-p2bs4\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.631356 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5cb978-dca7-4d06-8c84-c06588c6eacd-config-data\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.734908 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5cb978-dca7-4d06-8c84-c06588c6eacd-combined-ca-bundle\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.735153 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5cb978-dca7-4d06-8c84-c06588c6eacd-scripts\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.735269 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2bs4\" (UniqueName: \"kubernetes.io/projected/8c5cb978-dca7-4d06-8c84-c06588c6eacd-kube-api-access-p2bs4\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.735351 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5cb978-dca7-4d06-8c84-c06588c6eacd-config-data\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.735498 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5cb978-dca7-4d06-8c84-c06588c6eacd-logs\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.736491 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5cb978-dca7-4d06-8c84-c06588c6eacd-logs\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.740211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5cb978-dca7-4d06-8c84-c06588c6eacd-combined-ca-bundle\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.740739 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5cb978-dca7-4d06-8c84-c06588c6eacd-config-data\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.742500 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5cb978-dca7-4d06-8c84-c06588c6eacd-scripts\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.752957 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2bs4\" (UniqueName: \"kubernetes.io/projected/8c5cb978-dca7-4d06-8c84-c06588c6eacd-kube-api-access-p2bs4\") pod \"placement-5c8c7ff5f6-w4pdr\" (UID: \"8c5cb978-dca7-4d06-8c84-c06588c6eacd\") " pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:13 crc kubenswrapper[4945]: I1206 08:21:13.872529 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:14 crc kubenswrapper[4945]: I1206 08:21:14.319501 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5c8c7ff5f6-w4pdr"] Dec 06 08:21:14 crc kubenswrapper[4945]: I1206 08:21:14.473494 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c8c7ff5f6-w4pdr" event={"ID":"8c5cb978-dca7-4d06-8c84-c06588c6eacd","Type":"ContainerStarted","Data":"32ae086903a6cc2f7e1564f13745ea76a35bcce59350b68beb3c1ca79c8d6c20"} Dec 06 08:21:14 crc kubenswrapper[4945]: I1206 08:21:14.878588 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:21:14 crc kubenswrapper[4945]: I1206 08:21:14.937897 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759f69b595-gcn84"] Dec 06 08:21:14 crc kubenswrapper[4945]: I1206 08:21:14.938170 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-759f69b595-gcn84" podUID="aa27172f-213d-4a3a-b9df-6b6ad05de6b5" containerName="dnsmasq-dns" containerID="cri-o://baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7" gracePeriod=10 Dec 06 08:21:15 crc kubenswrapper[4945]: E1206 08:21:15.072858 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa27172f_213d_4a3a_b9df_6b6ad05de6b5.slice/crio-conmon-baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa27172f_213d_4a3a_b9df_6b6ad05de6b5.slice/crio-baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7.scope\": RecentStats: unable to find data in memory cache]" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.420383 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.481944 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c8c7ff5f6-w4pdr" event={"ID":"8c5cb978-dca7-4d06-8c84-c06588c6eacd","Type":"ContainerStarted","Data":"864c6a3c5d4a94991db529be3fdf317160f4536ee46a72183897ece822477b2f"} Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.481996 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c8c7ff5f6-w4pdr" event={"ID":"8c5cb978-dca7-4d06-8c84-c06588c6eacd","Type":"ContainerStarted","Data":"77a8c3c502dad935794deb1de30ae0a06f662caf6ab160e6b1f3698c524d3940"} Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.483124 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.483152 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.485208 4945 generic.go:334] "Generic (PLEG): container finished" podID="aa27172f-213d-4a3a-b9df-6b6ad05de6b5" containerID="baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7" exitCode=0 Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.485246 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759f69b595-gcn84" event={"ID":"aa27172f-213d-4a3a-b9df-6b6ad05de6b5","Type":"ContainerDied","Data":"baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7"} Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.485268 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759f69b595-gcn84" event={"ID":"aa27172f-213d-4a3a-b9df-6b6ad05de6b5","Type":"ContainerDied","Data":"f091699a809c174f4827ebaec20733180794bee775392b05d698dfdde63af797"} Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.485300 4945 scope.go:117] "RemoveContainer" containerID="baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.485424 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759f69b595-gcn84" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.508086 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5c8c7ff5f6-w4pdr" podStartSLOduration=2.508066774 podStartE2EDuration="2.508066774s" podCreationTimestamp="2025-12-06 08:21:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:21:15.49781424 +0000 UTC m=+5288.952675284" watchObservedRunningTime="2025-12-06 08:21:15.508066774 +0000 UTC m=+5288.962927818" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.509540 4945 scope.go:117] "RemoveContainer" containerID="0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.539884 4945 scope.go:117] "RemoveContainer" containerID="baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7" Dec 06 08:21:15 crc kubenswrapper[4945]: E1206 08:21:15.540432 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7\": container with ID starting with baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7 not found: ID does not exist" containerID="baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.540466 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7"} err="failed to get container status \"baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7\": rpc error: code = NotFound desc = could not find container \"baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7\": container with ID starting with baca0a7fd7cf42b2a10de26b2077a4601b681d98c29624761c770ee66e0e98a7 not found: ID does not exist" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.540488 4945 scope.go:117] "RemoveContainer" containerID="0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336" Dec 06 08:21:15 crc kubenswrapper[4945]: E1206 08:21:15.540837 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336\": container with ID starting with 0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336 not found: ID does not exist" containerID="0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.540861 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336"} err="failed to get container status \"0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336\": rpc error: code = NotFound desc = could not find container \"0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336\": container with ID starting with 0bdef30a039254e882e4f899778fcd1562c07e6f8992f418f8c045584b0c0336 not found: ID does not exist" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.575933 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-config\") pod \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.576009 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-sb\") pod \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.576032 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-dns-svc\") pod \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.576129 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-nb\") pod \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.576152 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4p8b\" (UniqueName: \"kubernetes.io/projected/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-kube-api-access-l4p8b\") pod \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\" (UID: \"aa27172f-213d-4a3a-b9df-6b6ad05de6b5\") " Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.600505 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-kube-api-access-l4p8b" (OuterVolumeSpecName: "kube-api-access-l4p8b") pod "aa27172f-213d-4a3a-b9df-6b6ad05de6b5" (UID: "aa27172f-213d-4a3a-b9df-6b6ad05de6b5"). InnerVolumeSpecName "kube-api-access-l4p8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.636075 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa27172f-213d-4a3a-b9df-6b6ad05de6b5" (UID: "aa27172f-213d-4a3a-b9df-6b6ad05de6b5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.650767 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-config" (OuterVolumeSpecName: "config") pod "aa27172f-213d-4a3a-b9df-6b6ad05de6b5" (UID: "aa27172f-213d-4a3a-b9df-6b6ad05de6b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.678584 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4p8b\" (UniqueName: \"kubernetes.io/projected/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-kube-api-access-l4p8b\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.678630 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.678646 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.690943 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa27172f-213d-4a3a-b9df-6b6ad05de6b5" (UID: "aa27172f-213d-4a3a-b9df-6b6ad05de6b5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.709828 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa27172f-213d-4a3a-b9df-6b6ad05de6b5" (UID: "aa27172f-213d-4a3a-b9df-6b6ad05de6b5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.780395 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.780421 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa27172f-213d-4a3a-b9df-6b6ad05de6b5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.820441 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759f69b595-gcn84"] Dec 06 08:21:15 crc kubenswrapper[4945]: I1206 08:21:15.831708 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-759f69b595-gcn84"] Dec 06 08:21:16 crc kubenswrapper[4945]: I1206 08:21:16.966613 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa27172f-213d-4a3a-b9df-6b6ad05de6b5" path="/var/lib/kubelet/pods/aa27172f-213d-4a3a-b9df-6b6ad05de6b5/volumes" Dec 06 08:21:24 crc kubenswrapper[4945]: I1206 08:21:24.800683 4945 scope.go:117] "RemoveContainer" containerID="428b11892998719a52f7cde2735cda8e80c06df12c4b80e3e4fdd817422035d6" Dec 06 08:21:24 crc kubenswrapper[4945]: I1206 08:21:24.956076 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:21:24 crc kubenswrapper[4945]: E1206 08:21:24.956295 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:21:39 crc kubenswrapper[4945]: I1206 08:21:39.953919 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:21:39 crc kubenswrapper[4945]: E1206 08:21:39.955684 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:21:44 crc kubenswrapper[4945]: I1206 08:21:44.849860 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:44 crc kubenswrapper[4945]: I1206 08:21:44.851963 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5c8c7ff5f6-w4pdr" Dec 06 08:21:54 crc kubenswrapper[4945]: I1206 08:21:54.953429 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:21:54 crc kubenswrapper[4945]: E1206 08:21:54.954294 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:22:07 crc kubenswrapper[4945]: I1206 08:22:07.953492 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:22:07 crc kubenswrapper[4945]: E1206 08:22:07.954191 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.298559 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-n2tgx"] Dec 06 08:22:09 crc kubenswrapper[4945]: E1206 08:22:09.299231 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa27172f-213d-4a3a-b9df-6b6ad05de6b5" containerName="init" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.299245 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa27172f-213d-4a3a-b9df-6b6ad05de6b5" containerName="init" Dec 06 08:22:09 crc kubenswrapper[4945]: E1206 08:22:09.299264 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa27172f-213d-4a3a-b9df-6b6ad05de6b5" containerName="dnsmasq-dns" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.299270 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa27172f-213d-4a3a-b9df-6b6ad05de6b5" containerName="dnsmasq-dns" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.299455 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa27172f-213d-4a3a-b9df-6b6ad05de6b5" containerName="dnsmasq-dns" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.300015 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-n2tgx" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.314545 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-n2tgx"] Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.323461 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77a80ef0-e94a-4986-834d-026c7ea1bf82-operator-scripts\") pod \"nova-api-db-create-n2tgx\" (UID: \"77a80ef0-e94a-4986-834d-026c7ea1bf82\") " pod="openstack/nova-api-db-create-n2tgx" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.323525 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lc8j\" (UniqueName: \"kubernetes.io/projected/77a80ef0-e94a-4986-834d-026c7ea1bf82-kube-api-access-9lc8j\") pod \"nova-api-db-create-n2tgx\" (UID: \"77a80ef0-e94a-4986-834d-026c7ea1bf82\") " pod="openstack/nova-api-db-create-n2tgx" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.397710 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-cbw8h"] Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.399244 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cbw8h" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.407188 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cbw8h"] Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.425110 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77a80ef0-e94a-4986-834d-026c7ea1bf82-operator-scripts\") pod \"nova-api-db-create-n2tgx\" (UID: \"77a80ef0-e94a-4986-834d-026c7ea1bf82\") " pod="openstack/nova-api-db-create-n2tgx" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.425249 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lc8j\" (UniqueName: \"kubernetes.io/projected/77a80ef0-e94a-4986-834d-026c7ea1bf82-kube-api-access-9lc8j\") pod \"nova-api-db-create-n2tgx\" (UID: \"77a80ef0-e94a-4986-834d-026c7ea1bf82\") " pod="openstack/nova-api-db-create-n2tgx" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.426249 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77a80ef0-e94a-4986-834d-026c7ea1bf82-operator-scripts\") pod \"nova-api-db-create-n2tgx\" (UID: \"77a80ef0-e94a-4986-834d-026c7ea1bf82\") " pod="openstack/nova-api-db-create-n2tgx" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.443435 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lc8j\" (UniqueName: \"kubernetes.io/projected/77a80ef0-e94a-4986-834d-026c7ea1bf82-kube-api-access-9lc8j\") pod \"nova-api-db-create-n2tgx\" (UID: \"77a80ef0-e94a-4986-834d-026c7ea1bf82\") " pod="openstack/nova-api-db-create-n2tgx" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.498665 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3cc2-account-create-update-sjmzc"] Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.500386 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3cc2-account-create-update-sjmzc" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.505190 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.509469 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3cc2-account-create-update-sjmzc"] Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.527216 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzds9\" (UniqueName: \"kubernetes.io/projected/96746661-4c5d-4339-990d-56f8fa99e2ad-kube-api-access-tzds9\") pod \"nova-api-3cc2-account-create-update-sjmzc\" (UID: \"96746661-4c5d-4339-990d-56f8fa99e2ad\") " pod="openstack/nova-api-3cc2-account-create-update-sjmzc" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.527310 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmff9\" (UniqueName: \"kubernetes.io/projected/fd91a488-ec3c-4918-a147-ade496239f9c-kube-api-access-qmff9\") pod \"nova-cell0-db-create-cbw8h\" (UID: \"fd91a488-ec3c-4918-a147-ade496239f9c\") " pod="openstack/nova-cell0-db-create-cbw8h" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.527395 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd91a488-ec3c-4918-a147-ade496239f9c-operator-scripts\") pod \"nova-cell0-db-create-cbw8h\" (UID: \"fd91a488-ec3c-4918-a147-ade496239f9c\") " pod="openstack/nova-cell0-db-create-cbw8h" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.527474 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96746661-4c5d-4339-990d-56f8fa99e2ad-operator-scripts\") pod \"nova-api-3cc2-account-create-update-sjmzc\" (UID: \"96746661-4c5d-4339-990d-56f8fa99e2ad\") " pod="openstack/nova-api-3cc2-account-create-update-sjmzc" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.598339 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-p7r95"] Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.599618 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p7r95" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.608271 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-p7r95"] Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.622933 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-n2tgx" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.628759 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmff9\" (UniqueName: \"kubernetes.io/projected/fd91a488-ec3c-4918-a147-ade496239f9c-kube-api-access-qmff9\") pod \"nova-cell0-db-create-cbw8h\" (UID: \"fd91a488-ec3c-4918-a147-ade496239f9c\") " pod="openstack/nova-cell0-db-create-cbw8h" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.629109 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd91a488-ec3c-4918-a147-ade496239f9c-operator-scripts\") pod \"nova-cell0-db-create-cbw8h\" (UID: \"fd91a488-ec3c-4918-a147-ade496239f9c\") " pod="openstack/nova-cell0-db-create-cbw8h" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.629165 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96746661-4c5d-4339-990d-56f8fa99e2ad-operator-scripts\") pod \"nova-api-3cc2-account-create-update-sjmzc\" (UID: \"96746661-4c5d-4339-990d-56f8fa99e2ad\") " pod="openstack/nova-api-3cc2-account-create-update-sjmzc" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.629226 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzds9\" (UniqueName: \"kubernetes.io/projected/96746661-4c5d-4339-990d-56f8fa99e2ad-kube-api-access-tzds9\") pod \"nova-api-3cc2-account-create-update-sjmzc\" (UID: \"96746661-4c5d-4339-990d-56f8fa99e2ad\") " pod="openstack/nova-api-3cc2-account-create-update-sjmzc" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.629840 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd91a488-ec3c-4918-a147-ade496239f9c-operator-scripts\") pod \"nova-cell0-db-create-cbw8h\" (UID: \"fd91a488-ec3c-4918-a147-ade496239f9c\") " pod="openstack/nova-cell0-db-create-cbw8h" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.630346 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96746661-4c5d-4339-990d-56f8fa99e2ad-operator-scripts\") pod \"nova-api-3cc2-account-create-update-sjmzc\" (UID: \"96746661-4c5d-4339-990d-56f8fa99e2ad\") " pod="openstack/nova-api-3cc2-account-create-update-sjmzc" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.645588 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzds9\" (UniqueName: \"kubernetes.io/projected/96746661-4c5d-4339-990d-56f8fa99e2ad-kube-api-access-tzds9\") pod \"nova-api-3cc2-account-create-update-sjmzc\" (UID: \"96746661-4c5d-4339-990d-56f8fa99e2ad\") " pod="openstack/nova-api-3cc2-account-create-update-sjmzc" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.646482 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmff9\" (UniqueName: \"kubernetes.io/projected/fd91a488-ec3c-4918-a147-ade496239f9c-kube-api-access-qmff9\") pod \"nova-cell0-db-create-cbw8h\" (UID: \"fd91a488-ec3c-4918-a147-ade496239f9c\") " pod="openstack/nova-cell0-db-create-cbw8h" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.708856 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-a73d-account-create-update-p8rn4"] Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.710017 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.714314 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.716499 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cbw8h" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.732925 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4nlx\" (UniqueName: \"kubernetes.io/projected/d3869f07-21d7-4706-9054-9bc3d2bb5818-kube-api-access-k4nlx\") pod \"nova-cell1-db-create-p7r95\" (UID: \"d3869f07-21d7-4706-9054-9bc3d2bb5818\") " pod="openstack/nova-cell1-db-create-p7r95" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.733004 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3869f07-21d7-4706-9054-9bc3d2bb5818-operator-scripts\") pod \"nova-cell1-db-create-p7r95\" (UID: \"d3869f07-21d7-4706-9054-9bc3d2bb5818\") " pod="openstack/nova-cell1-db-create-p7r95" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.748492 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a73d-account-create-update-p8rn4"] Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.817185 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3cc2-account-create-update-sjmzc" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.836529 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzf5k\" (UniqueName: \"kubernetes.io/projected/4a994594-0156-406c-943d-6e97aaf35488-kube-api-access-hzf5k\") pod \"nova-cell0-a73d-account-create-update-p8rn4\" (UID: \"4a994594-0156-406c-943d-6e97aaf35488\") " pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.836730 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a994594-0156-406c-943d-6e97aaf35488-operator-scripts\") pod \"nova-cell0-a73d-account-create-update-p8rn4\" (UID: \"4a994594-0156-406c-943d-6e97aaf35488\") " pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.836803 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4nlx\" (UniqueName: \"kubernetes.io/projected/d3869f07-21d7-4706-9054-9bc3d2bb5818-kube-api-access-k4nlx\") pod \"nova-cell1-db-create-p7r95\" (UID: \"d3869f07-21d7-4706-9054-9bc3d2bb5818\") " pod="openstack/nova-cell1-db-create-p7r95" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.836849 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3869f07-21d7-4706-9054-9bc3d2bb5818-operator-scripts\") pod \"nova-cell1-db-create-p7r95\" (UID: \"d3869f07-21d7-4706-9054-9bc3d2bb5818\") " pod="openstack/nova-cell1-db-create-p7r95" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.837630 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3869f07-21d7-4706-9054-9bc3d2bb5818-operator-scripts\") pod \"nova-cell1-db-create-p7r95\" (UID: \"d3869f07-21d7-4706-9054-9bc3d2bb5818\") " pod="openstack/nova-cell1-db-create-p7r95" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.886161 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4nlx\" (UniqueName: \"kubernetes.io/projected/d3869f07-21d7-4706-9054-9bc3d2bb5818-kube-api-access-k4nlx\") pod \"nova-cell1-db-create-p7r95\" (UID: \"d3869f07-21d7-4706-9054-9bc3d2bb5818\") " pod="openstack/nova-cell1-db-create-p7r95" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.923695 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p7r95" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.938803 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzf5k\" (UniqueName: \"kubernetes.io/projected/4a994594-0156-406c-943d-6e97aaf35488-kube-api-access-hzf5k\") pod \"nova-cell0-a73d-account-create-update-p8rn4\" (UID: \"4a994594-0156-406c-943d-6e97aaf35488\") " pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.945178 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a994594-0156-406c-943d-6e97aaf35488-operator-scripts\") pod \"nova-cell0-a73d-account-create-update-p8rn4\" (UID: \"4a994594-0156-406c-943d-6e97aaf35488\") " pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.948903 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a994594-0156-406c-943d-6e97aaf35488-operator-scripts\") pod \"nova-cell0-a73d-account-create-update-p8rn4\" (UID: \"4a994594-0156-406c-943d-6e97aaf35488\") " pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.940109 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-031f-account-create-update-gxfhm"] Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.950209 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-031f-account-create-update-gxfhm" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.959474 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.965797 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzf5k\" (UniqueName: \"kubernetes.io/projected/4a994594-0156-406c-943d-6e97aaf35488-kube-api-access-hzf5k\") pod \"nova-cell0-a73d-account-create-update-p8rn4\" (UID: \"4a994594-0156-406c-943d-6e97aaf35488\") " pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" Dec 06 08:22:09 crc kubenswrapper[4945]: I1206 08:22:09.991203 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-031f-account-create-update-gxfhm"] Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.049676 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zws7f\" (UniqueName: \"kubernetes.io/projected/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-kube-api-access-zws7f\") pod \"nova-cell1-031f-account-create-update-gxfhm\" (UID: \"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a\") " pod="openstack/nova-cell1-031f-account-create-update-gxfhm" Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.049854 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-operator-scripts\") pod \"nova-cell1-031f-account-create-update-gxfhm\" (UID: \"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a\") " pod="openstack/nova-cell1-031f-account-create-update-gxfhm" Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.075391 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.151327 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-operator-scripts\") pod \"nova-cell1-031f-account-create-update-gxfhm\" (UID: \"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a\") " pod="openstack/nova-cell1-031f-account-create-update-gxfhm" Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.151691 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zws7f\" (UniqueName: \"kubernetes.io/projected/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-kube-api-access-zws7f\") pod \"nova-cell1-031f-account-create-update-gxfhm\" (UID: \"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a\") " pod="openstack/nova-cell1-031f-account-create-update-gxfhm" Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.152451 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-operator-scripts\") pod \"nova-cell1-031f-account-create-update-gxfhm\" (UID: \"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a\") " pod="openstack/nova-cell1-031f-account-create-update-gxfhm" Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.170331 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zws7f\" (UniqueName: \"kubernetes.io/projected/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-kube-api-access-zws7f\") pod \"nova-cell1-031f-account-create-update-gxfhm\" (UID: \"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a\") " pod="openstack/nova-cell1-031f-account-create-update-gxfhm" Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.211166 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-n2tgx"] Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.283315 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-031f-account-create-update-gxfhm" Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.359868 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3cc2-account-create-update-sjmzc"] Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.371419 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cbw8h"] Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.532212 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-p7r95"] Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.650616 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a73d-account-create-update-p8rn4"] Dec 06 08:22:10 crc kubenswrapper[4945]: I1206 08:22:10.853715 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-031f-account-create-update-gxfhm"] Dec 06 08:22:10 crc kubenswrapper[4945]: W1206 08:22:10.858690 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9035d48b_9f2d_44de_b24a_55a5ab1d4c9a.slice/crio-6051c79b819b11857d6d65609aadddd3453375164cd0c8677b8775ed6b4a6535 WatchSource:0}: Error finding container 6051c79b819b11857d6d65609aadddd3453375164cd0c8677b8775ed6b4a6535: Status 404 returned error can't find the container with id 6051c79b819b11857d6d65609aadddd3453375164cd0c8677b8775ed6b4a6535 Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:10.999955 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-p7r95" event={"ID":"d3869f07-21d7-4706-9054-9bc3d2bb5818","Type":"ContainerStarted","Data":"de795a4d2fe7a5336e49879a5efe0c55825e43cb11cd494cf5e1449151e7fee1"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.000008 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-p7r95" event={"ID":"d3869f07-21d7-4706-9054-9bc3d2bb5818","Type":"ContainerStarted","Data":"ba2e9740f9ed267b85bd86c725cefa5dac28f81155004143e8270fd3e1ceee8b"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.002756 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3cc2-account-create-update-sjmzc" event={"ID":"96746661-4c5d-4339-990d-56f8fa99e2ad","Type":"ContainerStarted","Data":"9e47a6f6c052df5d7fa1ff7776516d5ba6c0ff5927dfed4d68462994c0dd93c7"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.002929 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3cc2-account-create-update-sjmzc" event={"ID":"96746661-4c5d-4339-990d-56f8fa99e2ad","Type":"ContainerStarted","Data":"8140ba06529f93f2f5e5e674d509ac4727cbd5238f4eb18aeba08855d2a640c9"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.024671 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-031f-account-create-update-gxfhm" event={"ID":"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a","Type":"ContainerStarted","Data":"6051c79b819b11857d6d65609aadddd3453375164cd0c8677b8775ed6b4a6535"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.031871 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cbw8h" event={"ID":"fd91a488-ec3c-4918-a147-ade496239f9c","Type":"ContainerStarted","Data":"b5ad938b22ccc2cf50ba0b26ca59b5f8d3b7530b9aa3147b2483c8c1a5d3e128"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.031915 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cbw8h" event={"ID":"fd91a488-ec3c-4918-a147-ade496239f9c","Type":"ContainerStarted","Data":"2c9a97183d715f850ded0433b138504aa77a4a866151c5af2dc26fc7297583c3"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.042037 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-3cc2-account-create-update-sjmzc" podStartSLOduration=2.042019471 podStartE2EDuration="2.042019471s" podCreationTimestamp="2025-12-06 08:22:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:22:11.033411741 +0000 UTC m=+5344.488272785" watchObservedRunningTime="2025-12-06 08:22:11.042019471 +0000 UTC m=+5344.496880515" Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.043851 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-p7r95" podStartSLOduration=2.0438447 podStartE2EDuration="2.0438447s" podCreationTimestamp="2025-12-06 08:22:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:22:11.018255945 +0000 UTC m=+5344.473117009" watchObservedRunningTime="2025-12-06 08:22:11.0438447 +0000 UTC m=+5344.498705744" Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.045506 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" event={"ID":"4a994594-0156-406c-943d-6e97aaf35488","Type":"ContainerStarted","Data":"59fb3a094a0c214eda54f96364742eae7f10f943d6cf12264386b024a3e8a9bf"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.045536 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" event={"ID":"4a994594-0156-406c-943d-6e97aaf35488","Type":"ContainerStarted","Data":"03cb2afa017a59bf3cd510f5678fd505a3f998d3724d0da3aa3920de33d857bb"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.048737 4945 generic.go:334] "Generic (PLEG): container finished" podID="77a80ef0-e94a-4986-834d-026c7ea1bf82" containerID="10384df17bdd621cd4c741a2622d092127bfe5be48248a0cab800ae7540058fc" exitCode=0 Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.048772 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-n2tgx" event={"ID":"77a80ef0-e94a-4986-834d-026c7ea1bf82","Type":"ContainerDied","Data":"10384df17bdd621cd4c741a2622d092127bfe5be48248a0cab800ae7540058fc"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.048792 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-n2tgx" event={"ID":"77a80ef0-e94a-4986-834d-026c7ea1bf82","Type":"ContainerStarted","Data":"7cfd54e1e1f635db43ff97124b53b2f7edd90418d9a4a529f690e58b7c360950"} Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.058765 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-cbw8h" podStartSLOduration=2.058745149 podStartE2EDuration="2.058745149s" podCreationTimestamp="2025-12-06 08:22:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:22:11.055488602 +0000 UTC m=+5344.510349676" watchObservedRunningTime="2025-12-06 08:22:11.058745149 +0000 UTC m=+5344.513606193" Dec 06 08:22:11 crc kubenswrapper[4945]: I1206 08:22:11.076412 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" podStartSLOduration=2.076391961 podStartE2EDuration="2.076391961s" podCreationTimestamp="2025-12-06 08:22:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:22:11.068154081 +0000 UTC m=+5344.523015125" watchObservedRunningTime="2025-12-06 08:22:11.076391961 +0000 UTC m=+5344.531253005" Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.058696 4945 generic.go:334] "Generic (PLEG): container finished" podID="9035d48b-9f2d-44de-b24a-55a5ab1d4c9a" containerID="b8f1e7a398b593df85f262ccbf846ca72ef60342e710fa576c434fbf1c3efa37" exitCode=0 Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.058751 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-031f-account-create-update-gxfhm" event={"ID":"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a","Type":"ContainerDied","Data":"b8f1e7a398b593df85f262ccbf846ca72ef60342e710fa576c434fbf1c3efa37"} Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.060567 4945 generic.go:334] "Generic (PLEG): container finished" podID="4a994594-0156-406c-943d-6e97aaf35488" containerID="59fb3a094a0c214eda54f96364742eae7f10f943d6cf12264386b024a3e8a9bf" exitCode=0 Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.060604 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" event={"ID":"4a994594-0156-406c-943d-6e97aaf35488","Type":"ContainerDied","Data":"59fb3a094a0c214eda54f96364742eae7f10f943d6cf12264386b024a3e8a9bf"} Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.062579 4945 generic.go:334] "Generic (PLEG): container finished" podID="fd91a488-ec3c-4918-a147-ade496239f9c" containerID="b5ad938b22ccc2cf50ba0b26ca59b5f8d3b7530b9aa3147b2483c8c1a5d3e128" exitCode=0 Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.062628 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cbw8h" event={"ID":"fd91a488-ec3c-4918-a147-ade496239f9c","Type":"ContainerDied","Data":"b5ad938b22ccc2cf50ba0b26ca59b5f8d3b7530b9aa3147b2483c8c1a5d3e128"} Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.064386 4945 generic.go:334] "Generic (PLEG): container finished" podID="d3869f07-21d7-4706-9054-9bc3d2bb5818" containerID="de795a4d2fe7a5336e49879a5efe0c55825e43cb11cd494cf5e1449151e7fee1" exitCode=0 Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.064513 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-p7r95" event={"ID":"d3869f07-21d7-4706-9054-9bc3d2bb5818","Type":"ContainerDied","Data":"de795a4d2fe7a5336e49879a5efe0c55825e43cb11cd494cf5e1449151e7fee1"} Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.066197 4945 generic.go:334] "Generic (PLEG): container finished" podID="96746661-4c5d-4339-990d-56f8fa99e2ad" containerID="9e47a6f6c052df5d7fa1ff7776516d5ba6c0ff5927dfed4d68462994c0dd93c7" exitCode=0 Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.066394 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3cc2-account-create-update-sjmzc" event={"ID":"96746661-4c5d-4339-990d-56f8fa99e2ad","Type":"ContainerDied","Data":"9e47a6f6c052df5d7fa1ff7776516d5ba6c0ff5927dfed4d68462994c0dd93c7"} Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.404617 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-n2tgx" Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.502533 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77a80ef0-e94a-4986-834d-026c7ea1bf82-operator-scripts\") pod \"77a80ef0-e94a-4986-834d-026c7ea1bf82\" (UID: \"77a80ef0-e94a-4986-834d-026c7ea1bf82\") " Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.502588 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lc8j\" (UniqueName: \"kubernetes.io/projected/77a80ef0-e94a-4986-834d-026c7ea1bf82-kube-api-access-9lc8j\") pod \"77a80ef0-e94a-4986-834d-026c7ea1bf82\" (UID: \"77a80ef0-e94a-4986-834d-026c7ea1bf82\") " Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.503576 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77a80ef0-e94a-4986-834d-026c7ea1bf82-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "77a80ef0-e94a-4986-834d-026c7ea1bf82" (UID: "77a80ef0-e94a-4986-834d-026c7ea1bf82"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.510584 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77a80ef0-e94a-4986-834d-026c7ea1bf82-kube-api-access-9lc8j" (OuterVolumeSpecName: "kube-api-access-9lc8j") pod "77a80ef0-e94a-4986-834d-026c7ea1bf82" (UID: "77a80ef0-e94a-4986-834d-026c7ea1bf82"). InnerVolumeSpecName "kube-api-access-9lc8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.606111 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lc8j\" (UniqueName: \"kubernetes.io/projected/77a80ef0-e94a-4986-834d-026c7ea1bf82-kube-api-access-9lc8j\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:12 crc kubenswrapper[4945]: I1206 08:22:12.606181 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77a80ef0-e94a-4986-834d-026c7ea1bf82-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.078801 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-n2tgx" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.079694 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-n2tgx" event={"ID":"77a80ef0-e94a-4986-834d-026c7ea1bf82","Type":"ContainerDied","Data":"7cfd54e1e1f635db43ff97124b53b2f7edd90418d9a4a529f690e58b7c360950"} Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.079721 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cfd54e1e1f635db43ff97124b53b2f7edd90418d9a4a529f690e58b7c360950" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.423097 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cbw8h" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.520518 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmff9\" (UniqueName: \"kubernetes.io/projected/fd91a488-ec3c-4918-a147-ade496239f9c-kube-api-access-qmff9\") pod \"fd91a488-ec3c-4918-a147-ade496239f9c\" (UID: \"fd91a488-ec3c-4918-a147-ade496239f9c\") " Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.520679 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd91a488-ec3c-4918-a147-ade496239f9c-operator-scripts\") pod \"fd91a488-ec3c-4918-a147-ade496239f9c\" (UID: \"fd91a488-ec3c-4918-a147-ade496239f9c\") " Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.521271 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd91a488-ec3c-4918-a147-ade496239f9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fd91a488-ec3c-4918-a147-ade496239f9c" (UID: "fd91a488-ec3c-4918-a147-ade496239f9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.524199 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd91a488-ec3c-4918-a147-ade496239f9c-kube-api-access-qmff9" (OuterVolumeSpecName: "kube-api-access-qmff9") pod "fd91a488-ec3c-4918-a147-ade496239f9c" (UID: "fd91a488-ec3c-4918-a147-ade496239f9c"). InnerVolumeSpecName "kube-api-access-qmff9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.587417 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-031f-account-create-update-gxfhm" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.594125 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p7r95" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.607710 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3cc2-account-create-update-sjmzc" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.623184 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3869f07-21d7-4706-9054-9bc3d2bb5818-operator-scripts\") pod \"d3869f07-21d7-4706-9054-9bc3d2bb5818\" (UID: \"d3869f07-21d7-4706-9054-9bc3d2bb5818\") " Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.623332 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4nlx\" (UniqueName: \"kubernetes.io/projected/d3869f07-21d7-4706-9054-9bc3d2bb5818-kube-api-access-k4nlx\") pod \"d3869f07-21d7-4706-9054-9bc3d2bb5818\" (UID: \"d3869f07-21d7-4706-9054-9bc3d2bb5818\") " Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.623461 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zws7f\" (UniqueName: \"kubernetes.io/projected/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-kube-api-access-zws7f\") pod \"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a\" (UID: \"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a\") " Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.623486 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-operator-scripts\") pod \"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a\" (UID: \"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a\") " Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.623830 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmff9\" (UniqueName: \"kubernetes.io/projected/fd91a488-ec3c-4918-a147-ade496239f9c-kube-api-access-qmff9\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.623846 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd91a488-ec3c-4918-a147-ade496239f9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.624760 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.624766 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9035d48b-9f2d-44de-b24a-55a5ab1d4c9a" (UID: "9035d48b-9f2d-44de-b24a-55a5ab1d4c9a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.625188 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3869f07-21d7-4706-9054-9bc3d2bb5818-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3869f07-21d7-4706-9054-9bc3d2bb5818" (UID: "d3869f07-21d7-4706-9054-9bc3d2bb5818"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.629837 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-kube-api-access-zws7f" (OuterVolumeSpecName: "kube-api-access-zws7f") pod "9035d48b-9f2d-44de-b24a-55a5ab1d4c9a" (UID: "9035d48b-9f2d-44de-b24a-55a5ab1d4c9a"). InnerVolumeSpecName "kube-api-access-zws7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.639829 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3869f07-21d7-4706-9054-9bc3d2bb5818-kube-api-access-k4nlx" (OuterVolumeSpecName: "kube-api-access-k4nlx") pod "d3869f07-21d7-4706-9054-9bc3d2bb5818" (UID: "d3869f07-21d7-4706-9054-9bc3d2bb5818"). InnerVolumeSpecName "kube-api-access-k4nlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.725370 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a994594-0156-406c-943d-6e97aaf35488-operator-scripts\") pod \"4a994594-0156-406c-943d-6e97aaf35488\" (UID: \"4a994594-0156-406c-943d-6e97aaf35488\") " Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.725519 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzds9\" (UniqueName: \"kubernetes.io/projected/96746661-4c5d-4339-990d-56f8fa99e2ad-kube-api-access-tzds9\") pod \"96746661-4c5d-4339-990d-56f8fa99e2ad\" (UID: \"96746661-4c5d-4339-990d-56f8fa99e2ad\") " Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.725663 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzf5k\" (UniqueName: \"kubernetes.io/projected/4a994594-0156-406c-943d-6e97aaf35488-kube-api-access-hzf5k\") pod \"4a994594-0156-406c-943d-6e97aaf35488\" (UID: \"4a994594-0156-406c-943d-6e97aaf35488\") " Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.726266 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96746661-4c5d-4339-990d-56f8fa99e2ad-operator-scripts\") pod \"96746661-4c5d-4339-990d-56f8fa99e2ad\" (UID: \"96746661-4c5d-4339-990d-56f8fa99e2ad\") " Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.726271 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a994594-0156-406c-943d-6e97aaf35488-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4a994594-0156-406c-943d-6e97aaf35488" (UID: "4a994594-0156-406c-943d-6e97aaf35488"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.726597 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96746661-4c5d-4339-990d-56f8fa99e2ad-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "96746661-4c5d-4339-990d-56f8fa99e2ad" (UID: "96746661-4c5d-4339-990d-56f8fa99e2ad"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.726958 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zws7f\" (UniqueName: \"kubernetes.io/projected/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-kube-api-access-zws7f\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.727028 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.727099 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96746661-4c5d-4339-990d-56f8fa99e2ad-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.727203 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3869f07-21d7-4706-9054-9bc3d2bb5818-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.727298 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a994594-0156-406c-943d-6e97aaf35488-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.727367 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4nlx\" (UniqueName: \"kubernetes.io/projected/d3869f07-21d7-4706-9054-9bc3d2bb5818-kube-api-access-k4nlx\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.728321 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96746661-4c5d-4339-990d-56f8fa99e2ad-kube-api-access-tzds9" (OuterVolumeSpecName: "kube-api-access-tzds9") pod "96746661-4c5d-4339-990d-56f8fa99e2ad" (UID: "96746661-4c5d-4339-990d-56f8fa99e2ad"). InnerVolumeSpecName "kube-api-access-tzds9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.728715 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a994594-0156-406c-943d-6e97aaf35488-kube-api-access-hzf5k" (OuterVolumeSpecName: "kube-api-access-hzf5k") pod "4a994594-0156-406c-943d-6e97aaf35488" (UID: "4a994594-0156-406c-943d-6e97aaf35488"). InnerVolumeSpecName "kube-api-access-hzf5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.829036 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzds9\" (UniqueName: \"kubernetes.io/projected/96746661-4c5d-4339-990d-56f8fa99e2ad-kube-api-access-tzds9\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:13 crc kubenswrapper[4945]: I1206 08:22:13.829095 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzf5k\" (UniqueName: \"kubernetes.io/projected/4a994594-0156-406c-943d-6e97aaf35488-kube-api-access-hzf5k\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.089449 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3cc2-account-create-update-sjmzc" event={"ID":"96746661-4c5d-4339-990d-56f8fa99e2ad","Type":"ContainerDied","Data":"8140ba06529f93f2f5e5e674d509ac4727cbd5238f4eb18aeba08855d2a640c9"} Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.089492 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8140ba06529f93f2f5e5e674d509ac4727cbd5238f4eb18aeba08855d2a640c9" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.089529 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3cc2-account-create-update-sjmzc" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.091917 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-031f-account-create-update-gxfhm" event={"ID":"9035d48b-9f2d-44de-b24a-55a5ab1d4c9a","Type":"ContainerDied","Data":"6051c79b819b11857d6d65609aadddd3453375164cd0c8677b8775ed6b4a6535"} Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.091951 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6051c79b819b11857d6d65609aadddd3453375164cd0c8677b8775ed6b4a6535" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.092033 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-031f-account-create-update-gxfhm" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.094086 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" event={"ID":"4a994594-0156-406c-943d-6e97aaf35488","Type":"ContainerDied","Data":"03cb2afa017a59bf3cd510f5678fd505a3f998d3724d0da3aa3920de33d857bb"} Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.094111 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03cb2afa017a59bf3cd510f5678fd505a3f998d3724d0da3aa3920de33d857bb" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.094089 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a73d-account-create-update-p8rn4" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.095593 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cbw8h" event={"ID":"fd91a488-ec3c-4918-a147-ade496239f9c","Type":"ContainerDied","Data":"2c9a97183d715f850ded0433b138504aa77a4a866151c5af2dc26fc7297583c3"} Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.095641 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c9a97183d715f850ded0433b138504aa77a4a866151c5af2dc26fc7297583c3" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.095726 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cbw8h" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.106514 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-p7r95" event={"ID":"d3869f07-21d7-4706-9054-9bc3d2bb5818","Type":"ContainerDied","Data":"ba2e9740f9ed267b85bd86c725cefa5dac28f81155004143e8270fd3e1ceee8b"} Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.106583 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba2e9740f9ed267b85bd86c725cefa5dac28f81155004143e8270fd3e1ceee8b" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.106677 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p7r95" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.903543 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ppmxf"] Dec 06 08:22:14 crc kubenswrapper[4945]: E1206 08:22:14.904093 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a80ef0-e94a-4986-834d-026c7ea1bf82" containerName="mariadb-database-create" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904115 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a80ef0-e94a-4986-834d-026c7ea1bf82" containerName="mariadb-database-create" Dec 06 08:22:14 crc kubenswrapper[4945]: E1206 08:22:14.904152 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd91a488-ec3c-4918-a147-ade496239f9c" containerName="mariadb-database-create" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904162 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd91a488-ec3c-4918-a147-ade496239f9c" containerName="mariadb-database-create" Dec 06 08:22:14 crc kubenswrapper[4945]: E1206 08:22:14.904183 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9035d48b-9f2d-44de-b24a-55a5ab1d4c9a" containerName="mariadb-account-create-update" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904193 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9035d48b-9f2d-44de-b24a-55a5ab1d4c9a" containerName="mariadb-account-create-update" Dec 06 08:22:14 crc kubenswrapper[4945]: E1206 08:22:14.904211 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96746661-4c5d-4339-990d-56f8fa99e2ad" containerName="mariadb-account-create-update" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904221 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="96746661-4c5d-4339-990d-56f8fa99e2ad" containerName="mariadb-account-create-update" Dec 06 08:22:14 crc kubenswrapper[4945]: E1206 08:22:14.904249 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3869f07-21d7-4706-9054-9bc3d2bb5818" containerName="mariadb-database-create" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904259 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3869f07-21d7-4706-9054-9bc3d2bb5818" containerName="mariadb-database-create" Dec 06 08:22:14 crc kubenswrapper[4945]: E1206 08:22:14.904300 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a994594-0156-406c-943d-6e97aaf35488" containerName="mariadb-account-create-update" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904312 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a994594-0156-406c-943d-6e97aaf35488" containerName="mariadb-account-create-update" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904551 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3869f07-21d7-4706-9054-9bc3d2bb5818" containerName="mariadb-database-create" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904570 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9035d48b-9f2d-44de-b24a-55a5ab1d4c9a" containerName="mariadb-account-create-update" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904590 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="96746661-4c5d-4339-990d-56f8fa99e2ad" containerName="mariadb-account-create-update" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904603 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd91a488-ec3c-4918-a147-ade496239f9c" containerName="mariadb-database-create" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904614 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="77a80ef0-e94a-4986-834d-026c7ea1bf82" containerName="mariadb-database-create" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.904629 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a994594-0156-406c-943d-6e97aaf35488" containerName="mariadb-account-create-update" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.905406 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.907632 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.907842 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.909927 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-8hmp8" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.916188 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ppmxf"] Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.950608 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-config-data\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.950726 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp7td\" (UniqueName: \"kubernetes.io/projected/804fd7d1-c742-43a9-9342-f792e005248d-kube-api-access-hp7td\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.950773 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:14 crc kubenswrapper[4945]: I1206 08:22:14.950835 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-scripts\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:15 crc kubenswrapper[4945]: I1206 08:22:15.052714 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp7td\" (UniqueName: \"kubernetes.io/projected/804fd7d1-c742-43a9-9342-f792e005248d-kube-api-access-hp7td\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:15 crc kubenswrapper[4945]: I1206 08:22:15.052803 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:15 crc kubenswrapper[4945]: I1206 08:22:15.052853 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-scripts\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:15 crc kubenswrapper[4945]: I1206 08:22:15.052999 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-config-data\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:15 crc kubenswrapper[4945]: I1206 08:22:15.057766 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-scripts\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:15 crc kubenswrapper[4945]: I1206 08:22:15.057997 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-config-data\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:15 crc kubenswrapper[4945]: I1206 08:22:15.058478 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:15 crc kubenswrapper[4945]: I1206 08:22:15.067391 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp7td\" (UniqueName: \"kubernetes.io/projected/804fd7d1-c742-43a9-9342-f792e005248d-kube-api-access-hp7td\") pod \"nova-cell0-conductor-db-sync-ppmxf\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:15 crc kubenswrapper[4945]: I1206 08:22:15.244602 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:15 crc kubenswrapper[4945]: I1206 08:22:15.755829 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ppmxf"] Dec 06 08:22:16 crc kubenswrapper[4945]: I1206 08:22:16.126375 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ppmxf" event={"ID":"804fd7d1-c742-43a9-9342-f792e005248d","Type":"ContainerStarted","Data":"246488490da9a6c9753dc4d7614cd7f5e2b4b3352f7229ec62006a65883ee105"} Dec 06 08:22:20 crc kubenswrapper[4945]: I1206 08:22:20.953954 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:22:20 crc kubenswrapper[4945]: E1206 08:22:20.955019 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:22:24 crc kubenswrapper[4945]: I1206 08:22:24.196182 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ppmxf" event={"ID":"804fd7d1-c742-43a9-9342-f792e005248d","Type":"ContainerStarted","Data":"7a2bd3e660d2a4969fe84d91f9fdfac941095f9f82a9feccd239f1787b60f0dc"} Dec 06 08:22:24 crc kubenswrapper[4945]: I1206 08:22:24.214772 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-ppmxf" podStartSLOduration=2.046254079 podStartE2EDuration="10.214756316s" podCreationTimestamp="2025-12-06 08:22:14 +0000 UTC" firstStartedPulling="2025-12-06 08:22:15.751494453 +0000 UTC m=+5349.206355497" lastFinishedPulling="2025-12-06 08:22:23.9199967 +0000 UTC m=+5357.374857734" observedRunningTime="2025-12-06 08:22:24.212023053 +0000 UTC m=+5357.666884097" watchObservedRunningTime="2025-12-06 08:22:24.214756316 +0000 UTC m=+5357.669617360" Dec 06 08:22:24 crc kubenswrapper[4945]: I1206 08:22:24.964209 4945 scope.go:117] "RemoveContainer" containerID="b91f7825d10936ccc7c298ec92cce4b1cb2c001b18565b7189aca2484035f00f" Dec 06 08:22:25 crc kubenswrapper[4945]: I1206 08:22:25.005530 4945 scope.go:117] "RemoveContainer" containerID="48c040a1e3360d33634d04e1fd88cddb12860a2ce18399f08577c03561062694" Dec 06 08:22:25 crc kubenswrapper[4945]: I1206 08:22:25.031117 4945 scope.go:117] "RemoveContainer" containerID="c9de46e2658c6071dc8376f6b0acc2d5183c3f17f8b3bd5b7a5efc36d51e8aba" Dec 06 08:22:29 crc kubenswrapper[4945]: I1206 08:22:29.237238 4945 generic.go:334] "Generic (PLEG): container finished" podID="804fd7d1-c742-43a9-9342-f792e005248d" containerID="7a2bd3e660d2a4969fe84d91f9fdfac941095f9f82a9feccd239f1787b60f0dc" exitCode=0 Dec 06 08:22:29 crc kubenswrapper[4945]: I1206 08:22:29.237357 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ppmxf" event={"ID":"804fd7d1-c742-43a9-9342-f792e005248d","Type":"ContainerDied","Data":"7a2bd3e660d2a4969fe84d91f9fdfac941095f9f82a9feccd239f1787b60f0dc"} Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.736373 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.846442 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-config-data\") pod \"804fd7d1-c742-43a9-9342-f792e005248d\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.846508 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp7td\" (UniqueName: \"kubernetes.io/projected/804fd7d1-c742-43a9-9342-f792e005248d-kube-api-access-hp7td\") pod \"804fd7d1-c742-43a9-9342-f792e005248d\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.846619 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-combined-ca-bundle\") pod \"804fd7d1-c742-43a9-9342-f792e005248d\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.846668 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-scripts\") pod \"804fd7d1-c742-43a9-9342-f792e005248d\" (UID: \"804fd7d1-c742-43a9-9342-f792e005248d\") " Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.851743 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-scripts" (OuterVolumeSpecName: "scripts") pod "804fd7d1-c742-43a9-9342-f792e005248d" (UID: "804fd7d1-c742-43a9-9342-f792e005248d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.852375 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/804fd7d1-c742-43a9-9342-f792e005248d-kube-api-access-hp7td" (OuterVolumeSpecName: "kube-api-access-hp7td") pod "804fd7d1-c742-43a9-9342-f792e005248d" (UID: "804fd7d1-c742-43a9-9342-f792e005248d"). InnerVolumeSpecName "kube-api-access-hp7td". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.874758 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-config-data" (OuterVolumeSpecName: "config-data") pod "804fd7d1-c742-43a9-9342-f792e005248d" (UID: "804fd7d1-c742-43a9-9342-f792e005248d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.877109 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "804fd7d1-c742-43a9-9342-f792e005248d" (UID: "804fd7d1-c742-43a9-9342-f792e005248d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.948215 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.948254 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp7td\" (UniqueName: \"kubernetes.io/projected/804fd7d1-c742-43a9-9342-f792e005248d-kube-api-access-hp7td\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.948271 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:30 crc kubenswrapper[4945]: I1206 08:22:30.948302 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/804fd7d1-c742-43a9-9342-f792e005248d-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.255082 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ppmxf" event={"ID":"804fd7d1-c742-43a9-9342-f792e005248d","Type":"ContainerDied","Data":"246488490da9a6c9753dc4d7614cd7f5e2b4b3352f7229ec62006a65883ee105"} Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.255128 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="246488490da9a6c9753dc4d7614cd7f5e2b4b3352f7229ec62006a65883ee105" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.255148 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ppmxf" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.331911 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 08:22:31 crc kubenswrapper[4945]: E1206 08:22:31.332397 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="804fd7d1-c742-43a9-9342-f792e005248d" containerName="nova-cell0-conductor-db-sync" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.332435 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="804fd7d1-c742-43a9-9342-f792e005248d" containerName="nova-cell0-conductor-db-sync" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.332691 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="804fd7d1-c742-43a9-9342-f792e005248d" containerName="nova-cell0-conductor-db-sync" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.333426 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.337728 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.348038 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-8hmp8" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.348967 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.367053 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcj7v\" (UniqueName: \"kubernetes.io/projected/741d99e3-a6c8-4156-bb83-2ced76434e51-kube-api-access-hcj7v\") pod \"nova-cell0-conductor-0\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.367149 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.367401 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.469065 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcj7v\" (UniqueName: \"kubernetes.io/projected/741d99e3-a6c8-4156-bb83-2ced76434e51-kube-api-access-hcj7v\") pod \"nova-cell0-conductor-0\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.469128 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.469239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.474215 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.474262 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.484746 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcj7v\" (UniqueName: \"kubernetes.io/projected/741d99e3-a6c8-4156-bb83-2ced76434e51-kube-api-access-hcj7v\") pod \"nova-cell0-conductor-0\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:31 crc kubenswrapper[4945]: I1206 08:22:31.655709 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:32 crc kubenswrapper[4945]: I1206 08:22:32.071654 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 08:22:32 crc kubenswrapper[4945]: I1206 08:22:32.264877 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"741d99e3-a6c8-4156-bb83-2ced76434e51","Type":"ContainerStarted","Data":"6444e684635d8d2694b5ac08930ed5a0516a0e6c23e16ab1e38889dc2e37db78"} Dec 06 08:22:33 crc kubenswrapper[4945]: I1206 08:22:33.276663 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"741d99e3-a6c8-4156-bb83-2ced76434e51","Type":"ContainerStarted","Data":"5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb"} Dec 06 08:22:33 crc kubenswrapper[4945]: I1206 08:22:33.279664 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:33 crc kubenswrapper[4945]: I1206 08:22:33.304318 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.3042981940000002 podStartE2EDuration="2.304298194s" podCreationTimestamp="2025-12-06 08:22:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:22:33.293033112 +0000 UTC m=+5366.747894156" watchObservedRunningTime="2025-12-06 08:22:33.304298194 +0000 UTC m=+5366.759159238" Dec 06 08:22:34 crc kubenswrapper[4945]: I1206 08:22:34.954045 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:22:34 crc kubenswrapper[4945]: E1206 08:22:34.954727 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:22:37 crc kubenswrapper[4945]: I1206 08:22:37.931899 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rxs54"] Dec 06 08:22:37 crc kubenswrapper[4945]: I1206 08:22:37.935036 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:37 crc kubenswrapper[4945]: I1206 08:22:37.939680 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rxs54"] Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.086942 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-utilities\") pod \"community-operators-rxs54\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.087259 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-catalog-content\") pod \"community-operators-rxs54\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.087329 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86gtx\" (UniqueName: \"kubernetes.io/projected/e137d224-e766-4063-bf00-85169d5adfa1-kube-api-access-86gtx\") pod \"community-operators-rxs54\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.190477 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-utilities\") pod \"community-operators-rxs54\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.190537 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-catalog-content\") pod \"community-operators-rxs54\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.190562 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86gtx\" (UniqueName: \"kubernetes.io/projected/e137d224-e766-4063-bf00-85169d5adfa1-kube-api-access-86gtx\") pod \"community-operators-rxs54\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.191459 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-utilities\") pod \"community-operators-rxs54\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.191684 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-catalog-content\") pod \"community-operators-rxs54\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.211981 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86gtx\" (UniqueName: \"kubernetes.io/projected/e137d224-e766-4063-bf00-85169d5adfa1-kube-api-access-86gtx\") pod \"community-operators-rxs54\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.265372 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:38 crc kubenswrapper[4945]: I1206 08:22:38.758691 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rxs54"] Dec 06 08:22:39 crc kubenswrapper[4945]: I1206 08:22:39.348605 4945 generic.go:334] "Generic (PLEG): container finished" podID="e137d224-e766-4063-bf00-85169d5adfa1" containerID="be18bb8ca8d46644a39216021cad96e1acf40ce78cf513156d37875548051980" exitCode=0 Dec 06 08:22:39 crc kubenswrapper[4945]: I1206 08:22:39.348649 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxs54" event={"ID":"e137d224-e766-4063-bf00-85169d5adfa1","Type":"ContainerDied","Data":"be18bb8ca8d46644a39216021cad96e1acf40ce78cf513156d37875548051980"} Dec 06 08:22:39 crc kubenswrapper[4945]: I1206 08:22:39.348931 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxs54" event={"ID":"e137d224-e766-4063-bf00-85169d5adfa1","Type":"ContainerStarted","Data":"fa259cc6ab913421e7afdc37d94bbe28eb5d40c7b11f5249d28640033cfe2038"} Dec 06 08:22:41 crc kubenswrapper[4945]: I1206 08:22:41.370102 4945 generic.go:334] "Generic (PLEG): container finished" podID="e137d224-e766-4063-bf00-85169d5adfa1" containerID="efe0f149abda91ea43289a33951acc28c5c2f268172dd9d2c8db95774adea28e" exitCode=0 Dec 06 08:22:41 crc kubenswrapper[4945]: I1206 08:22:41.370200 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxs54" event={"ID":"e137d224-e766-4063-bf00-85169d5adfa1","Type":"ContainerDied","Data":"efe0f149abda91ea43289a33951acc28c5c2f268172dd9d2c8db95774adea28e"} Dec 06 08:22:41 crc kubenswrapper[4945]: I1206 08:22:41.687709 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.110645 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-f59xz"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.112161 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.114645 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.114759 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.136102 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-f59xz"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.183100 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-scripts\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.183548 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7j5m\" (UniqueName: \"kubernetes.io/projected/67e6aa41-80ba-4a23-8fe9-591546bb4b04-kube-api-access-w7j5m\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.183652 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.183698 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-config-data\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.245364 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.246821 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.250119 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.263912 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.288433 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.288495 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-config-data\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.288540 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gmq2\" (UniqueName: \"kubernetes.io/projected/2679acb2-90dd-4a75-9e58-11b73546c8b7-kube-api-access-4gmq2\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.288623 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-scripts\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.288704 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-config-data\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.288732 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2679acb2-90dd-4a75-9e58-11b73546c8b7-logs\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.288777 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.288805 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7j5m\" (UniqueName: \"kubernetes.io/projected/67e6aa41-80ba-4a23-8fe9-591546bb4b04-kube-api-access-w7j5m\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.294947 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-scripts\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.296250 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-config-data\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.306937 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7j5m\" (UniqueName: \"kubernetes.io/projected/67e6aa41-80ba-4a23-8fe9-591546bb4b04-kube-api-access-w7j5m\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.315021 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-f59xz\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.356600 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.358586 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.366004 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.380264 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.390189 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-config-data\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.390233 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6fkf\" (UniqueName: \"kubernetes.io/projected/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-kube-api-access-r6fkf\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.390253 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2679acb2-90dd-4a75-9e58-11b73546c8b7-logs\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.390793 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.390852 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gmq2\" (UniqueName: \"kubernetes.io/projected/2679acb2-90dd-4a75-9e58-11b73546c8b7-kube-api-access-4gmq2\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.390898 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-logs\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.390925 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.390961 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-config-data\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.391312 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2679acb2-90dd-4a75-9e58-11b73546c8b7-logs\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.396097 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.413437 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-config-data\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.432560 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.433786 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.440863 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.443923 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gmq2\" (UniqueName: \"kubernetes.io/projected/2679acb2-90dd-4a75-9e58-11b73546c8b7-kube-api-access-4gmq2\") pod \"nova-api-0\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.445674 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.454458 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.467762 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7879cc8c79-sn2f4"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.469877 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498175 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498213 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vqdd\" (UniqueName: \"kubernetes.io/projected/72bcfa97-b158-4da4-ae32-8a24fba0062e-kube-api-access-5vqdd\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498237 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-config-data\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498258 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498355 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-dns-svc\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498430 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6fkf\" (UniqueName: \"kubernetes.io/projected/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-kube-api-access-r6fkf\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498473 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-config\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498518 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bqdd\" (UniqueName: \"kubernetes.io/projected/0412207c-21f7-4657-9213-ef4fba09c229-kube-api-access-2bqdd\") pod \"nova-cell1-novncproxy-0\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498548 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-sb\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498567 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-nb\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498599 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-logs\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.498625 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.502378 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-logs\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.524413 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-config-data\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.536916 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.542337 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7879cc8c79-sn2f4"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.551165 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.557097 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6fkf\" (UniqueName: \"kubernetes.io/projected/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-kube-api-access-r6fkf\") pod \"nova-metadata-0\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.564962 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.573835 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.574743 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.592945 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601069 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601110 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vqdd\" (UniqueName: \"kubernetes.io/projected/72bcfa97-b158-4da4-ae32-8a24fba0062e-kube-api-access-5vqdd\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601138 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " pod="openstack/nova-scheduler-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601157 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601184 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-dns-svc\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601242 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-config-data\") pod \"nova-scheduler-0\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " pod="openstack/nova-scheduler-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601295 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-config\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601333 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bqdd\" (UniqueName: \"kubernetes.io/projected/0412207c-21f7-4657-9213-ef4fba09c229-kube-api-access-2bqdd\") pod \"nova-cell1-novncproxy-0\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601352 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzngg\" (UniqueName: \"kubernetes.io/projected/0c9542d8-84fc-4151-b31f-f63401d22f45-kube-api-access-qzngg\") pod \"nova-scheduler-0\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " pod="openstack/nova-scheduler-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601385 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-nb\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.601399 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-sb\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.602463 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-sb\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.602693 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-config\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.603006 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-nb\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.603431 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-dns-svc\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.604891 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.642428 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.647816 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bqdd\" (UniqueName: \"kubernetes.io/projected/0412207c-21f7-4657-9213-ef4fba09c229-kube-api-access-2bqdd\") pod \"nova-cell1-novncproxy-0\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.656203 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vqdd\" (UniqueName: \"kubernetes.io/projected/72bcfa97-b158-4da4-ae32-8a24fba0062e-kube-api-access-5vqdd\") pod \"dnsmasq-dns-7879cc8c79-sn2f4\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.705215 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-config-data\") pod \"nova-scheduler-0\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " pod="openstack/nova-scheduler-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.705326 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzngg\" (UniqueName: \"kubernetes.io/projected/0c9542d8-84fc-4151-b31f-f63401d22f45-kube-api-access-qzngg\") pod \"nova-scheduler-0\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " pod="openstack/nova-scheduler-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.705406 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " pod="openstack/nova-scheduler-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.737197 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-config-data\") pod \"nova-scheduler-0\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " pod="openstack/nova-scheduler-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.740983 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzngg\" (UniqueName: \"kubernetes.io/projected/0c9542d8-84fc-4151-b31f-f63401d22f45-kube-api-access-qzngg\") pod \"nova-scheduler-0\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " pod="openstack/nova-scheduler-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.745886 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " pod="openstack/nova-scheduler-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.781742 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.849347 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.866792 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:42 crc kubenswrapper[4945]: I1206 08:22:42.947298 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.176689 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.275684 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-f59xz"] Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.391838 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f59xz" event={"ID":"67e6aa41-80ba-4a23-8fe9-591546bb4b04","Type":"ContainerStarted","Data":"574a160fbe0859a70c7f1f58df536127e8490fbe35cb046e6ba1162828f31107"} Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.400664 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2679acb2-90dd-4a75-9e58-11b73546c8b7","Type":"ContainerStarted","Data":"46323ce8653603a391d552ea922301f00688737d7fb50ecbcaae09edbf41ae99"} Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.403712 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxs54" event={"ID":"e137d224-e766-4063-bf00-85169d5adfa1","Type":"ContainerStarted","Data":"0100f599b33389ed1eedbb084b1780aea65a0f140292b5a2e8714bbbe48f1bcc"} Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.432137 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rxs54" podStartSLOduration=2.678650249 podStartE2EDuration="6.432118541s" podCreationTimestamp="2025-12-06 08:22:37 +0000 UTC" firstStartedPulling="2025-12-06 08:22:39.350230422 +0000 UTC m=+5372.805091466" lastFinishedPulling="2025-12-06 08:22:43.103698724 +0000 UTC m=+5376.558559758" observedRunningTime="2025-12-06 08:22:43.426214653 +0000 UTC m=+5376.881075697" watchObservedRunningTime="2025-12-06 08:22:43.432118541 +0000 UTC m=+5376.886979585" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.456122 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pvdks"] Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.457863 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.459892 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.460320 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.491762 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pvdks"] Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.520606 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gv92\" (UniqueName: \"kubernetes.io/projected/bff2045b-07d6-45fa-93a5-578ea063487d-kube-api-access-4gv92\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.520670 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.520720 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-config-data\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.520757 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-scripts\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.525034 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:22:43 crc kubenswrapper[4945]: W1206 08:22:43.551695 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0412207c_21f7_4657_9213_ef4fba09c229.slice/crio-d89055514ad5ebf818362cae9782a764e36cedeae7531a0223b5396c1a896fa1 WatchSource:0}: Error finding container d89055514ad5ebf818362cae9782a764e36cedeae7531a0223b5396c1a896fa1: Status 404 returned error can't find the container with id d89055514ad5ebf818362cae9782a764e36cedeae7531a0223b5396c1a896fa1 Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.553980 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7879cc8c79-sn2f4"] Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.564031 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.615508 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.622971 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-config-data\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.623040 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-scripts\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.623148 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gv92\" (UniqueName: \"kubernetes.io/projected/bff2045b-07d6-45fa-93a5-578ea063487d-kube-api-access-4gv92\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.623364 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.629090 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-config-data\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.631351 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.631397 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-scripts\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: W1206 08:22:43.635986 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c9542d8_84fc_4151_b31f_f63401d22f45.slice/crio-7d2885ef55a52458ff3f8fc02ef92c450e60752e28131589523f8562adc15c09 WatchSource:0}: Error finding container 7d2885ef55a52458ff3f8fc02ef92c450e60752e28131589523f8562adc15c09: Status 404 returned error can't find the container with id 7d2885ef55a52458ff3f8fc02ef92c450e60752e28131589523f8562adc15c09 Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.641161 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gv92\" (UniqueName: \"kubernetes.io/projected/bff2045b-07d6-45fa-93a5-578ea063487d-kube-api-access-4gv92\") pod \"nova-cell1-conductor-db-sync-pvdks\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:43 crc kubenswrapper[4945]: I1206 08:22:43.646654 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:44 crc kubenswrapper[4945]: I1206 08:22:44.185686 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pvdks"] Dec 06 08:22:44 crc kubenswrapper[4945]: W1206 08:22:44.196104 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbff2045b_07d6_45fa_93a5_578ea063487d.slice/crio-48eecb4816716dcbaf2f6e01376d2d9162546ea35717b8761fd379dfdf8a93f2 WatchSource:0}: Error finding container 48eecb4816716dcbaf2f6e01376d2d9162546ea35717b8761fd379dfdf8a93f2: Status 404 returned error can't find the container with id 48eecb4816716dcbaf2f6e01376d2d9162546ea35717b8761fd379dfdf8a93f2 Dec 06 08:22:44 crc kubenswrapper[4945]: I1206 08:22:44.415098 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0c9542d8-84fc-4151-b31f-f63401d22f45","Type":"ContainerStarted","Data":"7d2885ef55a52458ff3f8fc02ef92c450e60752e28131589523f8562adc15c09"} Dec 06 08:22:44 crc kubenswrapper[4945]: I1206 08:22:44.417139 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pvdks" event={"ID":"bff2045b-07d6-45fa-93a5-578ea063487d","Type":"ContainerStarted","Data":"48eecb4816716dcbaf2f6e01376d2d9162546ea35717b8761fd379dfdf8a93f2"} Dec 06 08:22:44 crc kubenswrapper[4945]: I1206 08:22:44.419973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5","Type":"ContainerStarted","Data":"6c1c46ff44610271960e938ae3a05b7ab5641e4178484f284296dfd4838d7fa6"} Dec 06 08:22:44 crc kubenswrapper[4945]: I1206 08:22:44.422057 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f59xz" event={"ID":"67e6aa41-80ba-4a23-8fe9-591546bb4b04","Type":"ContainerStarted","Data":"956ce91cf576767e9b6191b62ed776e5c2f90d5bef68b0555b671358c0911ae4"} Dec 06 08:22:44 crc kubenswrapper[4945]: I1206 08:22:44.425323 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0412207c-21f7-4657-9213-ef4fba09c229","Type":"ContainerStarted","Data":"d89055514ad5ebf818362cae9782a764e36cedeae7531a0223b5396c1a896fa1"} Dec 06 08:22:44 crc kubenswrapper[4945]: I1206 08:22:44.428815 4945 generic.go:334] "Generic (PLEG): container finished" podID="72bcfa97-b158-4da4-ae32-8a24fba0062e" containerID="ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49" exitCode=0 Dec 06 08:22:44 crc kubenswrapper[4945]: I1206 08:22:44.429079 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" event={"ID":"72bcfa97-b158-4da4-ae32-8a24fba0062e","Type":"ContainerDied","Data":"ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49"} Dec 06 08:22:44 crc kubenswrapper[4945]: I1206 08:22:44.429118 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" event={"ID":"72bcfa97-b158-4da4-ae32-8a24fba0062e","Type":"ContainerStarted","Data":"23dadb8a4f5580ffbeae633266ae4088e885b0dc60d3b642c5197848635de704"} Dec 06 08:22:44 crc kubenswrapper[4945]: I1206 08:22:44.452300 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-f59xz" podStartSLOduration=2.452257495 podStartE2EDuration="2.452257495s" podCreationTimestamp="2025-12-06 08:22:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:22:44.440956192 +0000 UTC m=+5377.895817236" watchObservedRunningTime="2025-12-06 08:22:44.452257495 +0000 UTC m=+5377.907118539" Dec 06 08:22:45 crc kubenswrapper[4945]: I1206 08:22:45.440426 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pvdks" event={"ID":"bff2045b-07d6-45fa-93a5-578ea063487d","Type":"ContainerStarted","Data":"61960e1e0cb8d6380227bae539a7086f39ca9c9894bc3ee625a151d39e45fc3e"} Dec 06 08:22:45 crc kubenswrapper[4945]: I1206 08:22:45.453370 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" event={"ID":"72bcfa97-b158-4da4-ae32-8a24fba0062e","Type":"ContainerStarted","Data":"e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c"} Dec 06 08:22:45 crc kubenswrapper[4945]: I1206 08:22:45.461710 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-pvdks" podStartSLOduration=2.461689212 podStartE2EDuration="2.461689212s" podCreationTimestamp="2025-12-06 08:22:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:22:45.460206992 +0000 UTC m=+5378.915068046" watchObservedRunningTime="2025-12-06 08:22:45.461689212 +0000 UTC m=+5378.916550256" Dec 06 08:22:45 crc kubenswrapper[4945]: I1206 08:22:45.480914 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" podStartSLOduration=3.480897746 podStartE2EDuration="3.480897746s" podCreationTimestamp="2025-12-06 08:22:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:22:45.477292419 +0000 UTC m=+5378.932153463" watchObservedRunningTime="2025-12-06 08:22:45.480897746 +0000 UTC m=+5378.935758790" Dec 06 08:22:46 crc kubenswrapper[4945]: I1206 08:22:46.467664 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.477776 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5","Type":"ContainerStarted","Data":"87b0503ec21cc0fbc782d92f278fc59e73b6e4d5c1d298b33759be629693b934"} Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.478420 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5","Type":"ContainerStarted","Data":"8e5a0b39093a858d9e61ce13472fc4694f96c3f4924cf5eb95fedf79183f458d"} Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.480090 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0412207c-21f7-4657-9213-ef4fba09c229","Type":"ContainerStarted","Data":"cedab13422fa411a54f0aa69ed9f44eab4068035dd24099951417aeba4b4c39b"} Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.482743 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2679acb2-90dd-4a75-9e58-11b73546c8b7","Type":"ContainerStarted","Data":"0fe59e13cfb36e50dc432633cf59bcc80a8085cea3e1f31ab4126749b7bb1bf3"} Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.482778 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2679acb2-90dd-4a75-9e58-11b73546c8b7","Type":"ContainerStarted","Data":"728d73092c351b4538cb7d2ff96259565fe4ec68a6ebd9130ad54352397b0700"} Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.484152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0c9542d8-84fc-4151-b31f-f63401d22f45","Type":"ContainerStarted","Data":"26fef944cc251b2e6754a6f0194737b0ac451a6d906cec1e350dfd07f0f300e3"} Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.485653 4945 generic.go:334] "Generic (PLEG): container finished" podID="bff2045b-07d6-45fa-93a5-578ea063487d" containerID="61960e1e0cb8d6380227bae539a7086f39ca9c9894bc3ee625a151d39e45fc3e" exitCode=0 Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.485692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pvdks" event={"ID":"bff2045b-07d6-45fa-93a5-578ea063487d","Type":"ContainerDied","Data":"61960e1e0cb8d6380227bae539a7086f39ca9c9894bc3ee625a151d39e45fc3e"} Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.504977 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.672670011 podStartE2EDuration="5.504959249s" podCreationTimestamp="2025-12-06 08:22:42 +0000 UTC" firstStartedPulling="2025-12-06 08:22:43.494293124 +0000 UTC m=+5376.949154168" lastFinishedPulling="2025-12-06 08:22:46.326582352 +0000 UTC m=+5379.781443406" observedRunningTime="2025-12-06 08:22:47.500186331 +0000 UTC m=+5380.955047385" watchObservedRunningTime="2025-12-06 08:22:47.504959249 +0000 UTC m=+5380.959820293" Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.548039 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.777210578 podStartE2EDuration="5.548020861s" podCreationTimestamp="2025-12-06 08:22:42 +0000 UTC" firstStartedPulling="2025-12-06 08:22:43.555940054 +0000 UTC m=+5377.010801098" lastFinishedPulling="2025-12-06 08:22:46.326750337 +0000 UTC m=+5379.781611381" observedRunningTime="2025-12-06 08:22:47.539155214 +0000 UTC m=+5380.994016268" watchObservedRunningTime="2025-12-06 08:22:47.548020861 +0000 UTC m=+5381.002881905" Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.564580 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.410545737 podStartE2EDuration="5.564556343s" podCreationTimestamp="2025-12-06 08:22:42 +0000 UTC" firstStartedPulling="2025-12-06 08:22:43.178610758 +0000 UTC m=+5376.633471802" lastFinishedPulling="2025-12-06 08:22:46.332621364 +0000 UTC m=+5379.787482408" observedRunningTime="2025-12-06 08:22:47.560007421 +0000 UTC m=+5381.014868465" watchObservedRunningTime="2025-12-06 08:22:47.564556343 +0000 UTC m=+5381.019417397" Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.592809 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.907597246 podStartE2EDuration="5.592786668s" podCreationTimestamp="2025-12-06 08:22:42 +0000 UTC" firstStartedPulling="2025-12-06 08:22:43.640219699 +0000 UTC m=+5377.095080743" lastFinishedPulling="2025-12-06 08:22:46.325409131 +0000 UTC m=+5379.780270165" observedRunningTime="2025-12-06 08:22:47.578236719 +0000 UTC m=+5381.033097783" watchObservedRunningTime="2025-12-06 08:22:47.592786668 +0000 UTC m=+5381.047647712" Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.783345 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.783634 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.850918 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:47 crc kubenswrapper[4945]: I1206 08:22:47.947685 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.265628 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.265925 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.857707 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.940184 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-scripts\") pod \"bff2045b-07d6-45fa-93a5-578ea063487d\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.940604 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-config-data\") pod \"bff2045b-07d6-45fa-93a5-578ea063487d\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.940652 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gv92\" (UniqueName: \"kubernetes.io/projected/bff2045b-07d6-45fa-93a5-578ea063487d-kube-api-access-4gv92\") pod \"bff2045b-07d6-45fa-93a5-578ea063487d\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.940709 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-combined-ca-bundle\") pod \"bff2045b-07d6-45fa-93a5-578ea063487d\" (UID: \"bff2045b-07d6-45fa-93a5-578ea063487d\") " Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.947414 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-scripts" (OuterVolumeSpecName: "scripts") pod "bff2045b-07d6-45fa-93a5-578ea063487d" (UID: "bff2045b-07d6-45fa-93a5-578ea063487d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.947600 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bff2045b-07d6-45fa-93a5-578ea063487d-kube-api-access-4gv92" (OuterVolumeSpecName: "kube-api-access-4gv92") pod "bff2045b-07d6-45fa-93a5-578ea063487d" (UID: "bff2045b-07d6-45fa-93a5-578ea063487d"). InnerVolumeSpecName "kube-api-access-4gv92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.953445 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:22:48 crc kubenswrapper[4945]: E1206 08:22:48.953723 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.974448 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bff2045b-07d6-45fa-93a5-578ea063487d" (UID: "bff2045b-07d6-45fa-93a5-578ea063487d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:22:48 crc kubenswrapper[4945]: I1206 08:22:48.977642 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-config-data" (OuterVolumeSpecName: "config-data") pod "bff2045b-07d6-45fa-93a5-578ea063487d" (UID: "bff2045b-07d6-45fa-93a5-578ea063487d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.042902 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.042945 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.042958 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gv92\" (UniqueName: \"kubernetes.io/projected/bff2045b-07d6-45fa-93a5-578ea063487d-kube-api-access-4gv92\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.042970 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff2045b-07d6-45fa-93a5-578ea063487d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.425456 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rxs54" podUID="e137d224-e766-4063-bf00-85169d5adfa1" containerName="registry-server" probeResult="failure" output=< Dec 06 08:22:49 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 08:22:49 crc kubenswrapper[4945]: > Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.505775 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pvdks" event={"ID":"bff2045b-07d6-45fa-93a5-578ea063487d","Type":"ContainerDied","Data":"48eecb4816716dcbaf2f6e01376d2d9162546ea35717b8761fd379dfdf8a93f2"} Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.505821 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pvdks" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.505835 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48eecb4816716dcbaf2f6e01376d2d9162546ea35717b8761fd379dfdf8a93f2" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.599817 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 08:22:49 crc kubenswrapper[4945]: E1206 08:22:49.600409 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bff2045b-07d6-45fa-93a5-578ea063487d" containerName="nova-cell1-conductor-db-sync" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.600437 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bff2045b-07d6-45fa-93a5-578ea063487d" containerName="nova-cell1-conductor-db-sync" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.601330 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bff2045b-07d6-45fa-93a5-578ea063487d" containerName="nova-cell1-conductor-db-sync" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.602197 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.605480 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.610340 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.756578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.756661 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbpl2\" (UniqueName: \"kubernetes.io/projected/3896b278-e2a5-4576-a764-06cf62902109-kube-api-access-xbpl2\") pod \"nova-cell1-conductor-0\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.756761 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.858043 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbpl2\" (UniqueName: \"kubernetes.io/projected/3896b278-e2a5-4576-a764-06cf62902109-kube-api-access-xbpl2\") pod \"nova-cell1-conductor-0\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.858168 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.858221 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.862100 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.862638 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.888872 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbpl2\" (UniqueName: \"kubernetes.io/projected/3896b278-e2a5-4576-a764-06cf62902109-kube-api-access-xbpl2\") pod \"nova-cell1-conductor-0\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:49 crc kubenswrapper[4945]: I1206 08:22:49.958915 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:50 crc kubenswrapper[4945]: I1206 08:22:50.416364 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 08:22:50 crc kubenswrapper[4945]: W1206 08:22:50.424139 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3896b278_e2a5_4576_a764_06cf62902109.slice/crio-0c963f5b0a7ab016bd769f8a1b43eb3e84f2092dc305e2df4aae08c172bc0a4c WatchSource:0}: Error finding container 0c963f5b0a7ab016bd769f8a1b43eb3e84f2092dc305e2df4aae08c172bc0a4c: Status 404 returned error can't find the container with id 0c963f5b0a7ab016bd769f8a1b43eb3e84f2092dc305e2df4aae08c172bc0a4c Dec 06 08:22:50 crc kubenswrapper[4945]: I1206 08:22:50.519592 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"3896b278-e2a5-4576-a764-06cf62902109","Type":"ContainerStarted","Data":"0c963f5b0a7ab016bd769f8a1b43eb3e84f2092dc305e2df4aae08c172bc0a4c"} Dec 06 08:22:51 crc kubenswrapper[4945]: I1206 08:22:51.527799 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"3896b278-e2a5-4576-a764-06cf62902109","Type":"ContainerStarted","Data":"4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf"} Dec 06 08:22:51 crc kubenswrapper[4945]: I1206 08:22:51.528132 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 06 08:22:51 crc kubenswrapper[4945]: I1206 08:22:51.529899 4945 generic.go:334] "Generic (PLEG): container finished" podID="67e6aa41-80ba-4a23-8fe9-591546bb4b04" containerID="956ce91cf576767e9b6191b62ed776e5c2f90d5bef68b0555b671358c0911ae4" exitCode=0 Dec 06 08:22:51 crc kubenswrapper[4945]: I1206 08:22:51.529927 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f59xz" event={"ID":"67e6aa41-80ba-4a23-8fe9-591546bb4b04","Type":"ContainerDied","Data":"956ce91cf576767e9b6191b62ed776e5c2f90d5bef68b0555b671358c0911ae4"} Dec 06 08:22:51 crc kubenswrapper[4945]: I1206 08:22:51.545049 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.545019782 podStartE2EDuration="2.545019782s" podCreationTimestamp="2025-12-06 08:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:22:51.541975551 +0000 UTC m=+5384.996836595" watchObservedRunningTime="2025-12-06 08:22:51.545019782 +0000 UTC m=+5384.999880826" Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.575326 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.575386 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.783489 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.784053 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.851448 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.862821 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.868458 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.937034 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.947575 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.952957 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b5d65bdf9-jv9rg"] Dec 06 08:22:52 crc kubenswrapper[4945]: I1206 08:22:52.953215 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" podUID="9d508820-8bdf-4efe-9adc-d40c565b4378" containerName="dnsmasq-dns" containerID="cri-o://bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5" gracePeriod=10 Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.008914 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.066864 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-scripts\") pod \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.066981 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-combined-ca-bundle\") pod \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.067017 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-config-data\") pod \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.067086 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7j5m\" (UniqueName: \"kubernetes.io/projected/67e6aa41-80ba-4a23-8fe9-591546bb4b04-kube-api-access-w7j5m\") pod \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\" (UID: \"67e6aa41-80ba-4a23-8fe9-591546bb4b04\") " Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.087937 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e6aa41-80ba-4a23-8fe9-591546bb4b04-kube-api-access-w7j5m" (OuterVolumeSpecName: "kube-api-access-w7j5m") pod "67e6aa41-80ba-4a23-8fe9-591546bb4b04" (UID: "67e6aa41-80ba-4a23-8fe9-591546bb4b04"). InnerVolumeSpecName "kube-api-access-w7j5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.120775 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-scripts" (OuterVolumeSpecName: "scripts") pod "67e6aa41-80ba-4a23-8fe9-591546bb4b04" (UID: "67e6aa41-80ba-4a23-8fe9-591546bb4b04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.142874 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-config-data" (OuterVolumeSpecName: "config-data") pod "67e6aa41-80ba-4a23-8fe9-591546bb4b04" (UID: "67e6aa41-80ba-4a23-8fe9-591546bb4b04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.172101 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.172130 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.172140 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7j5m\" (UniqueName: \"kubernetes.io/projected/67e6aa41-80ba-4a23-8fe9-591546bb4b04-kube-api-access-w7j5m\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.228918 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67e6aa41-80ba-4a23-8fe9-591546bb4b04" (UID: "67e6aa41-80ba-4a23-8fe9-591546bb4b04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.275942 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e6aa41-80ba-4a23-8fe9-591546bb4b04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.505322 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.573884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f59xz" event={"ID":"67e6aa41-80ba-4a23-8fe9-591546bb4b04","Type":"ContainerDied","Data":"574a160fbe0859a70c7f1f58df536127e8490fbe35cb046e6ba1162828f31107"} Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.573932 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="574a160fbe0859a70c7f1f58df536127e8490fbe35cb046e6ba1162828f31107" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.574010 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f59xz" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.582259 4945 generic.go:334] "Generic (PLEG): container finished" podID="9d508820-8bdf-4efe-9adc-d40c565b4378" containerID="bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5" exitCode=0 Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.582339 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" event={"ID":"9d508820-8bdf-4efe-9adc-d40c565b4378","Type":"ContainerDied","Data":"bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5"} Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.582407 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" event={"ID":"9d508820-8bdf-4efe-9adc-d40c565b4378","Type":"ContainerDied","Data":"62b898cdb2444c4fb453b1c71e9f4d6628427725d4fac57f2fb84bbe7c62976a"} Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.582427 4945 scope.go:117] "RemoveContainer" containerID="bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.582428 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b5d65bdf9-jv9rg" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.596101 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-sb\") pod \"9d508820-8bdf-4efe-9adc-d40c565b4378\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.596266 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-config\") pod \"9d508820-8bdf-4efe-9adc-d40c565b4378\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.596342 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc64z\" (UniqueName: \"kubernetes.io/projected/9d508820-8bdf-4efe-9adc-d40c565b4378-kube-api-access-gc64z\") pod \"9d508820-8bdf-4efe-9adc-d40c565b4378\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.596373 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-nb\") pod \"9d508820-8bdf-4efe-9adc-d40c565b4378\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.596492 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-dns-svc\") pod \"9d508820-8bdf-4efe-9adc-d40c565b4378\" (UID: \"9d508820-8bdf-4efe-9adc-d40c565b4378\") " Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.613533 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d508820-8bdf-4efe-9adc-d40c565b4378-kube-api-access-gc64z" (OuterVolumeSpecName: "kube-api-access-gc64z") pod "9d508820-8bdf-4efe-9adc-d40c565b4378" (UID: "9d508820-8bdf-4efe-9adc-d40c565b4378"). InnerVolumeSpecName "kube-api-access-gc64z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.639780 4945 scope.go:117] "RemoveContainer" containerID="068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.653398 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.669826 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.58:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.670438 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.58:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.686427 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9d508820-8bdf-4efe-9adc-d40c565b4378" (UID: "9d508820-8bdf-4efe-9adc-d40c565b4378"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.690643 4945 scope.go:117] "RemoveContainer" containerID="bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5" Dec 06 08:22:53 crc kubenswrapper[4945]: E1206 08:22:53.691173 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5\": container with ID starting with bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5 not found: ID does not exist" containerID="bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.691200 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5"} err="failed to get container status \"bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5\": rpc error: code = NotFound desc = could not find container \"bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5\": container with ID starting with bafaf46e6829006e464654dd7938e4b479809b63908d2982a8c9c7f6d63e43e5 not found: ID does not exist" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.691223 4945 scope.go:117] "RemoveContainer" containerID="068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7" Dec 06 08:22:53 crc kubenswrapper[4945]: E1206 08:22:53.691444 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7\": container with ID starting with 068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7 not found: ID does not exist" containerID="068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.691463 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7"} err="failed to get container status \"068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7\": rpc error: code = NotFound desc = could not find container \"068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7\": container with ID starting with 068b8f9dcd187ac24eddfa0b0cdc126900b276eac2e7fcf828a45d33e0b558b7 not found: ID does not exist" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.695986 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9d508820-8bdf-4efe-9adc-d40c565b4378" (UID: "9d508820-8bdf-4efe-9adc-d40c565b4378"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.698138 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc64z\" (UniqueName: \"kubernetes.io/projected/9d508820-8bdf-4efe-9adc-d40c565b4378-kube-api-access-gc64z\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.698160 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.698168 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.708224 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9d508820-8bdf-4efe-9adc-d40c565b4378" (UID: "9d508820-8bdf-4efe-9adc-d40c565b4378"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.715008 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-config" (OuterVolumeSpecName: "config") pod "9d508820-8bdf-4efe-9adc-d40c565b4378" (UID: "9d508820-8bdf-4efe-9adc-d40c565b4378"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.730413 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.764343 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.764651 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerName="nova-api-log" containerID="cri-o://728d73092c351b4538cb7d2ff96259565fe4ec68a6ebd9130ad54352397b0700" gracePeriod=30 Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.765072 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerName="nova-api-api" containerID="cri-o://0fe59e13cfb36e50dc432633cf59bcc80a8085cea3e1f31ab4126749b7bb1bf3" gracePeriod=30 Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.799454 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.799496 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d508820-8bdf-4efe-9adc-d40c565b4378-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.827408 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.827641 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerName="nova-metadata-log" containerID="cri-o://8e5a0b39093a858d9e61ce13472fc4694f96c3f4924cf5eb95fedf79183f458d" gracePeriod=30 Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.828115 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerName="nova-metadata-metadata" containerID="cri-o://87b0503ec21cc0fbc782d92f278fc59e73b6e4d5c1d298b33759be629693b934" gracePeriod=30 Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.838114 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.59:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.838636 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.59:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.943953 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b5d65bdf9-jv9rg"] Dec 06 08:22:53 crc kubenswrapper[4945]: I1206 08:22:53.952403 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b5d65bdf9-jv9rg"] Dec 06 08:22:54 crc kubenswrapper[4945]: I1206 08:22:54.218067 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:22:54 crc kubenswrapper[4945]: I1206 08:22:54.593461 4945 generic.go:334] "Generic (PLEG): container finished" podID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerID="728d73092c351b4538cb7d2ff96259565fe4ec68a6ebd9130ad54352397b0700" exitCode=143 Dec 06 08:22:54 crc kubenswrapper[4945]: I1206 08:22:54.593528 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2679acb2-90dd-4a75-9e58-11b73546c8b7","Type":"ContainerDied","Data":"728d73092c351b4538cb7d2ff96259565fe4ec68a6ebd9130ad54352397b0700"} Dec 06 08:22:54 crc kubenswrapper[4945]: I1206 08:22:54.600218 4945 generic.go:334] "Generic (PLEG): container finished" podID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerID="8e5a0b39093a858d9e61ce13472fc4694f96c3f4924cf5eb95fedf79183f458d" exitCode=143 Dec 06 08:22:54 crc kubenswrapper[4945]: I1206 08:22:54.600314 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5","Type":"ContainerDied","Data":"8e5a0b39093a858d9e61ce13472fc4694f96c3f4924cf5eb95fedf79183f458d"} Dec 06 08:22:54 crc kubenswrapper[4945]: I1206 08:22:54.969045 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d508820-8bdf-4efe-9adc-d40c565b4378" path="/var/lib/kubelet/pods/9d508820-8bdf-4efe-9adc-d40c565b4378/volumes" Dec 06 08:22:55 crc kubenswrapper[4945]: I1206 08:22:55.608611 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0c9542d8-84fc-4151-b31f-f63401d22f45" containerName="nova-scheduler-scheduler" containerID="cri-o://26fef944cc251b2e6754a6f0194737b0ac451a6d906cec1e350dfd07f0f300e3" gracePeriod=30 Dec 06 08:22:57 crc kubenswrapper[4945]: E1206 08:22:57.949481 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="26fef944cc251b2e6754a6f0194737b0ac451a6d906cec1e350dfd07f0f300e3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:22:57 crc kubenswrapper[4945]: E1206 08:22:57.951322 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="26fef944cc251b2e6754a6f0194737b0ac451a6d906cec1e350dfd07f0f300e3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:22:57 crc kubenswrapper[4945]: E1206 08:22:57.952349 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="26fef944cc251b2e6754a6f0194737b0ac451a6d906cec1e350dfd07f0f300e3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:22:57 crc kubenswrapper[4945]: E1206 08:22:57.952386 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0c9542d8-84fc-4151-b31f-f63401d22f45" containerName="nova-scheduler-scheduler" Dec 06 08:22:58 crc kubenswrapper[4945]: I1206 08:22:58.309435 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:58 crc kubenswrapper[4945]: I1206 08:22:58.359405 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:22:59 crc kubenswrapper[4945]: I1206 08:22:59.521802 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rxs54"] Dec 06 08:22:59 crc kubenswrapper[4945]: I1206 08:22:59.652445 4945 generic.go:334] "Generic (PLEG): container finished" podID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerID="0fe59e13cfb36e50dc432633cf59bcc80a8085cea3e1f31ab4126749b7bb1bf3" exitCode=0 Dec 06 08:22:59 crc kubenswrapper[4945]: I1206 08:22:59.652542 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2679acb2-90dd-4a75-9e58-11b73546c8b7","Type":"ContainerDied","Data":"0fe59e13cfb36e50dc432633cf59bcc80a8085cea3e1f31ab4126749b7bb1bf3"} Dec 06 08:22:59 crc kubenswrapper[4945]: I1206 08:22:59.653119 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rxs54" podUID="e137d224-e766-4063-bf00-85169d5adfa1" containerName="registry-server" containerID="cri-o://0100f599b33389ed1eedbb084b1780aea65a0f140292b5a2e8714bbbe48f1bcc" gracePeriod=2 Dec 06 08:22:59 crc kubenswrapper[4945]: I1206 08:22:59.982824 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:22:59 crc kubenswrapper[4945]: E1206 08:22:59.983186 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.036211 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.293937 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.327225 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-config-data\") pod \"2679acb2-90dd-4a75-9e58-11b73546c8b7\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.327469 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-combined-ca-bundle\") pod \"2679acb2-90dd-4a75-9e58-11b73546c8b7\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.327619 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gmq2\" (UniqueName: \"kubernetes.io/projected/2679acb2-90dd-4a75-9e58-11b73546c8b7-kube-api-access-4gmq2\") pod \"2679acb2-90dd-4a75-9e58-11b73546c8b7\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.327671 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2679acb2-90dd-4a75-9e58-11b73546c8b7-logs\") pod \"2679acb2-90dd-4a75-9e58-11b73546c8b7\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.328410 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2679acb2-90dd-4a75-9e58-11b73546c8b7-logs" (OuterVolumeSpecName: "logs") pod "2679acb2-90dd-4a75-9e58-11b73546c8b7" (UID: "2679acb2-90dd-4a75-9e58-11b73546c8b7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.334541 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2679acb2-90dd-4a75-9e58-11b73546c8b7-kube-api-access-4gmq2" (OuterVolumeSpecName: "kube-api-access-4gmq2") pod "2679acb2-90dd-4a75-9e58-11b73546c8b7" (UID: "2679acb2-90dd-4a75-9e58-11b73546c8b7"). InnerVolumeSpecName "kube-api-access-4gmq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:00 crc kubenswrapper[4945]: E1206 08:23:00.354218 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-combined-ca-bundle podName:2679acb2-90dd-4a75-9e58-11b73546c8b7 nodeName:}" failed. No retries permitted until 2025-12-06 08:23:00.854183879 +0000 UTC m=+5394.309044933 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-combined-ca-bundle") pod "2679acb2-90dd-4a75-9e58-11b73546c8b7" (UID: "2679acb2-90dd-4a75-9e58-11b73546c8b7") : error deleting /var/lib/kubelet/pods/2679acb2-90dd-4a75-9e58-11b73546c8b7/volume-subpaths: remove /var/lib/kubelet/pods/2679acb2-90dd-4a75-9e58-11b73546c8b7/volume-subpaths: no such file or directory Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.356694 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-config-data" (OuterVolumeSpecName: "config-data") pod "2679acb2-90dd-4a75-9e58-11b73546c8b7" (UID: "2679acb2-90dd-4a75-9e58-11b73546c8b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.429748 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gmq2\" (UniqueName: \"kubernetes.io/projected/2679acb2-90dd-4a75-9e58-11b73546c8b7-kube-api-access-4gmq2\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.429784 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2679acb2-90dd-4a75-9e58-11b73546c8b7-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.429795 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.511769 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-kj84k"] Dec 06 08:23:00 crc kubenswrapper[4945]: E1206 08:23:00.512180 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d508820-8bdf-4efe-9adc-d40c565b4378" containerName="dnsmasq-dns" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.512199 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d508820-8bdf-4efe-9adc-d40c565b4378" containerName="dnsmasq-dns" Dec 06 08:23:00 crc kubenswrapper[4945]: E1206 08:23:00.512212 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e6aa41-80ba-4a23-8fe9-591546bb4b04" containerName="nova-manage" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.512218 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e6aa41-80ba-4a23-8fe9-591546bb4b04" containerName="nova-manage" Dec 06 08:23:00 crc kubenswrapper[4945]: E1206 08:23:00.512234 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerName="nova-api-log" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.512250 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerName="nova-api-log" Dec 06 08:23:00 crc kubenswrapper[4945]: E1206 08:23:00.512286 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d508820-8bdf-4efe-9adc-d40c565b4378" containerName="init" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.512292 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d508820-8bdf-4efe-9adc-d40c565b4378" containerName="init" Dec 06 08:23:00 crc kubenswrapper[4945]: E1206 08:23:00.512304 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerName="nova-api-api" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.512310 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerName="nova-api-api" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.512500 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d508820-8bdf-4efe-9adc-d40c565b4378" containerName="dnsmasq-dns" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.512515 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerName="nova-api-log" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.512525 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" containerName="nova-api-api" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.512535 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e6aa41-80ba-4a23-8fe9-591546bb4b04" containerName="nova-manage" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.513288 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.515359 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.517968 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.521481 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kj84k"] Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.531870 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb2jn\" (UniqueName: \"kubernetes.io/projected/a133569e-416b-46a2-ab24-c1cfe9d4718d-kube-api-access-zb2jn\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.531944 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-config-data\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.532002 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.532022 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-scripts\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.632552 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.632601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-scripts\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.632726 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb2jn\" (UniqueName: \"kubernetes.io/projected/a133569e-416b-46a2-ab24-c1cfe9d4718d-kube-api-access-zb2jn\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.632775 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-config-data\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.637389 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.637399 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-config-data\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.639470 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-scripts\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.656051 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb2jn\" (UniqueName: \"kubernetes.io/projected/a133569e-416b-46a2-ab24-c1cfe9d4718d-kube-api-access-zb2jn\") pod \"nova-cell1-cell-mapping-kj84k\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.675410 4945 generic.go:334] "Generic (PLEG): container finished" podID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerID="87b0503ec21cc0fbc782d92f278fc59e73b6e4d5c1d298b33759be629693b934" exitCode=0 Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.675501 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5","Type":"ContainerDied","Data":"87b0503ec21cc0fbc782d92f278fc59e73b6e4d5c1d298b33759be629693b934"} Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.689543 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2679acb2-90dd-4a75-9e58-11b73546c8b7","Type":"ContainerDied","Data":"46323ce8653603a391d552ea922301f00688737d7fb50ecbcaae09edbf41ae99"} Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.689601 4945 scope.go:117] "RemoveContainer" containerID="0fe59e13cfb36e50dc432633cf59bcc80a8085cea3e1f31ab4126749b7bb1bf3" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.689558 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.698965 4945 generic.go:334] "Generic (PLEG): container finished" podID="e137d224-e766-4063-bf00-85169d5adfa1" containerID="0100f599b33389ed1eedbb084b1780aea65a0f140292b5a2e8714bbbe48f1bcc" exitCode=0 Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.699027 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxs54" event={"ID":"e137d224-e766-4063-bf00-85169d5adfa1","Type":"ContainerDied","Data":"0100f599b33389ed1eedbb084b1780aea65a0f140292b5a2e8714bbbe48f1bcc"} Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.702581 4945 generic.go:334] "Generic (PLEG): container finished" podID="0c9542d8-84fc-4151-b31f-f63401d22f45" containerID="26fef944cc251b2e6754a6f0194737b0ac451a6d906cec1e350dfd07f0f300e3" exitCode=0 Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.702622 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0c9542d8-84fc-4151-b31f-f63401d22f45","Type":"ContainerDied","Data":"26fef944cc251b2e6754a6f0194737b0ac451a6d906cec1e350dfd07f0f300e3"} Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.735223 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.751779 4945 scope.go:117] "RemoveContainer" containerID="728d73092c351b4538cb7d2ff96259565fe4ec68a6ebd9130ad54352397b0700" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.776531 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.846496 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.946856 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-combined-ca-bundle\") pod \"2679acb2-90dd-4a75-9e58-11b73546c8b7\" (UID: \"2679acb2-90dd-4a75-9e58-11b73546c8b7\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.946901 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86gtx\" (UniqueName: \"kubernetes.io/projected/e137d224-e766-4063-bf00-85169d5adfa1-kube-api-access-86gtx\") pod \"e137d224-e766-4063-bf00-85169d5adfa1\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.946977 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-utilities\") pod \"e137d224-e766-4063-bf00-85169d5adfa1\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.947014 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6fkf\" (UniqueName: \"kubernetes.io/projected/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-kube-api-access-r6fkf\") pod \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.947092 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-combined-ca-bundle\") pod \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.947132 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-logs\") pod \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.947153 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-catalog-content\") pod \"e137d224-e766-4063-bf00-85169d5adfa1\" (UID: \"e137d224-e766-4063-bf00-85169d5adfa1\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.947170 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-config-data\") pod \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\" (UID: \"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5\") " Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.950394 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-logs" (OuterVolumeSpecName: "logs") pod "1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" (UID: "1bc9d35a-31d2-4fbf-a605-08825bb7c0c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.957900 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-utilities" (OuterVolumeSpecName: "utilities") pod "e137d224-e766-4063-bf00-85169d5adfa1" (UID: "e137d224-e766-4063-bf00-85169d5adfa1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.983784 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e137d224-e766-4063-bf00-85169d5adfa1-kube-api-access-86gtx" (OuterVolumeSpecName: "kube-api-access-86gtx") pod "e137d224-e766-4063-bf00-85169d5adfa1" (UID: "e137d224-e766-4063-bf00-85169d5adfa1"). InnerVolumeSpecName "kube-api-access-86gtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.987071 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-config-data" (OuterVolumeSpecName: "config-data") pod "1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" (UID: "1bc9d35a-31d2-4fbf-a605-08825bb7c0c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:00 crc kubenswrapper[4945]: I1206 08:23:00.993051 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2679acb2-90dd-4a75-9e58-11b73546c8b7" (UID: "2679acb2-90dd-4a75-9e58-11b73546c8b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.001266 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-kube-api-access-r6fkf" (OuterVolumeSpecName: "kube-api-access-r6fkf") pod "1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" (UID: "1bc9d35a-31d2-4fbf-a605-08825bb7c0c5"). InnerVolumeSpecName "kube-api-access-r6fkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.004387 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" (UID: "1bc9d35a-31d2-4fbf-a605-08825bb7c0c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.017365 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e137d224-e766-4063-bf00-85169d5adfa1" (UID: "e137d224-e766-4063-bf00-85169d5adfa1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.050986 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2679acb2-90dd-4a75-9e58-11b73546c8b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.051026 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86gtx\" (UniqueName: \"kubernetes.io/projected/e137d224-e766-4063-bf00-85169d5adfa1-kube-api-access-86gtx\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.051036 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.051045 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6fkf\" (UniqueName: \"kubernetes.io/projected/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-kube-api-access-r6fkf\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.051053 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.051062 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.051070 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e137d224-e766-4063-bf00-85169d5adfa1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.051078 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.213502 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.334425 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.344196 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.360647 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: E1206 08:23:01.361179 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e137d224-e766-4063-bf00-85169d5adfa1" containerName="extract-content" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.361204 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e137d224-e766-4063-bf00-85169d5adfa1" containerName="extract-content" Dec 06 08:23:01 crc kubenswrapper[4945]: E1206 08:23:01.361224 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e137d224-e766-4063-bf00-85169d5adfa1" containerName="registry-server" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.361233 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e137d224-e766-4063-bf00-85169d5adfa1" containerName="registry-server" Dec 06 08:23:01 crc kubenswrapper[4945]: E1206 08:23:01.361255 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e137d224-e766-4063-bf00-85169d5adfa1" containerName="extract-utilities" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.361264 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e137d224-e766-4063-bf00-85169d5adfa1" containerName="extract-utilities" Dec 06 08:23:01 crc kubenswrapper[4945]: E1206 08:23:01.361301 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerName="nova-metadata-log" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.361309 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerName="nova-metadata-log" Dec 06 08:23:01 crc kubenswrapper[4945]: E1206 08:23:01.361338 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerName="nova-metadata-metadata" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.361346 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerName="nova-metadata-metadata" Dec 06 08:23:01 crc kubenswrapper[4945]: E1206 08:23:01.361357 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9542d8-84fc-4151-b31f-f63401d22f45" containerName="nova-scheduler-scheduler" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.361365 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9542d8-84fc-4151-b31f-f63401d22f45" containerName="nova-scheduler-scheduler" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.361551 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e137d224-e766-4063-bf00-85169d5adfa1" containerName="registry-server" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.361585 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerName="nova-metadata-log" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.361603 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" containerName="nova-metadata-metadata" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.361611 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9542d8-84fc-4151-b31f-f63401d22f45" containerName="nova-scheduler-scheduler" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.362777 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.365992 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.368880 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-combined-ca-bundle\") pod \"0c9542d8-84fc-4151-b31f-f63401d22f45\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.369211 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzngg\" (UniqueName: \"kubernetes.io/projected/0c9542d8-84fc-4151-b31f-f63401d22f45-kube-api-access-qzngg\") pod \"0c9542d8-84fc-4151-b31f-f63401d22f45\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.369628 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-config-data\") pod \"0c9542d8-84fc-4151-b31f-f63401d22f45\" (UID: \"0c9542d8-84fc-4151-b31f-f63401d22f45\") " Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.369820 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-config-data\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.369847 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.369882 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-logs\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.369915 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4kfz\" (UniqueName: \"kubernetes.io/projected/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-kube-api-access-k4kfz\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.375608 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.389461 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9542d8-84fc-4151-b31f-f63401d22f45-kube-api-access-qzngg" (OuterVolumeSpecName: "kube-api-access-qzngg") pod "0c9542d8-84fc-4151-b31f-f63401d22f45" (UID: "0c9542d8-84fc-4151-b31f-f63401d22f45"). InnerVolumeSpecName "kube-api-access-qzngg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.395867 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c9542d8-84fc-4151-b31f-f63401d22f45" (UID: "0c9542d8-84fc-4151-b31f-f63401d22f45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.399856 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-config-data" (OuterVolumeSpecName: "config-data") pod "0c9542d8-84fc-4151-b31f-f63401d22f45" (UID: "0c9542d8-84fc-4151-b31f-f63401d22f45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.445141 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kj84k"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.484214 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-config-data\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.484293 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.484361 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-logs\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.484420 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4kfz\" (UniqueName: \"kubernetes.io/projected/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-kube-api-access-k4kfz\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.484606 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.484624 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzngg\" (UniqueName: \"kubernetes.io/projected/0c9542d8-84fc-4151-b31f-f63401d22f45-kube-api-access-qzngg\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.484636 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9542d8-84fc-4151-b31f-f63401d22f45-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.485187 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-logs\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.487812 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.487938 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-config-data\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.503637 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4kfz\" (UniqueName: \"kubernetes.io/projected/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-kube-api-access-k4kfz\") pod \"nova-api-0\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.585464 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.718187 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kj84k" event={"ID":"a133569e-416b-46a2-ab24-c1cfe9d4718d","Type":"ContainerStarted","Data":"d764b626224cc9f20ff1c6240e8386ae706bb83d4cef60d7c3aa322dc369c78e"} Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.718466 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kj84k" event={"ID":"a133569e-416b-46a2-ab24-c1cfe9d4718d","Type":"ContainerStarted","Data":"be7f99edc2056b20f8850f1d8c677e6a2ec52613ee423798f925702c6016c5c4"} Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.721001 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0c9542d8-84fc-4151-b31f-f63401d22f45","Type":"ContainerDied","Data":"7d2885ef55a52458ff3f8fc02ef92c450e60752e28131589523f8562adc15c09"} Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.721041 4945 scope.go:117] "RemoveContainer" containerID="26fef944cc251b2e6754a6f0194737b0ac451a6d906cec1e350dfd07f0f300e3" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.721052 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.723408 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1bc9d35a-31d2-4fbf-a605-08825bb7c0c5","Type":"ContainerDied","Data":"6c1c46ff44610271960e938ae3a05b7ab5641e4178484f284296dfd4838d7fa6"} Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.723497 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.730313 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxs54" event={"ID":"e137d224-e766-4063-bf00-85169d5adfa1","Type":"ContainerDied","Data":"fa259cc6ab913421e7afdc37d94bbe28eb5d40c7b11f5249d28640033cfe2038"} Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.730415 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rxs54" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.741417 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-kj84k" podStartSLOduration=1.7413965839999999 podStartE2EDuration="1.741396584s" podCreationTimestamp="2025-12-06 08:23:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:23:01.735852005 +0000 UTC m=+5395.190713059" watchObservedRunningTime="2025-12-06 08:23:01.741396584 +0000 UTC m=+5395.196257638" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.770115 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.803931 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.810391 4945 scope.go:117] "RemoveContainer" containerID="87b0503ec21cc0fbc782d92f278fc59e73b6e4d5c1d298b33759be629693b934" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.821614 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.823466 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.826986 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.832042 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.840949 4945 scope.go:117] "RemoveContainer" containerID="8e5a0b39093a858d9e61ce13472fc4694f96c3f4924cf5eb95fedf79183f458d" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.841239 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.852821 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.863370 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rxs54"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.868702 4945 scope.go:117] "RemoveContainer" containerID="0100f599b33389ed1eedbb084b1780aea65a0f140292b5a2e8714bbbe48f1bcc" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.873457 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rxs54"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.885900 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.887520 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.890026 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.893762 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-config-data\") pod \"nova-scheduler-0\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.893809 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292f9a97-273c-431e-9c47-c0890d07c4ce-logs\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.893897 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.893946 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.893969 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-config-data\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.894000 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfzqw\" (UniqueName: \"kubernetes.io/projected/330175a1-2b61-4ad3-b170-9ee626141d7c-kube-api-access-rfzqw\") pod \"nova-scheduler-0\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.894080 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxq2j\" (UniqueName: \"kubernetes.io/projected/292f9a97-273c-431e-9c47-c0890d07c4ce-kube-api-access-nxq2j\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.894330 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.899346 4945 scope.go:117] "RemoveContainer" containerID="efe0f149abda91ea43289a33951acc28c5c2f268172dd9d2c8db95774adea28e" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.939150 4945 scope.go:117] "RemoveContainer" containerID="be18bb8ca8d46644a39216021cad96e1acf40ce78cf513156d37875548051980" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.994650 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxq2j\" (UniqueName: \"kubernetes.io/projected/292f9a97-273c-431e-9c47-c0890d07c4ce-kube-api-access-nxq2j\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.994697 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-config-data\") pod \"nova-scheduler-0\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.994731 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292f9a97-273c-431e-9c47-c0890d07c4ce-logs\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.994778 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.994808 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.994829 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-config-data\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.994858 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfzqw\" (UniqueName: \"kubernetes.io/projected/330175a1-2b61-4ad3-b170-9ee626141d7c-kube-api-access-rfzqw\") pod \"nova-scheduler-0\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.997488 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292f9a97-273c-431e-9c47-c0890d07c4ce-logs\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.999119 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-config-data\") pod \"nova-scheduler-0\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.999170 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-config-data\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.999644 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:01 crc kubenswrapper[4945]: I1206 08:23:01.999788 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.009911 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxq2j\" (UniqueName: \"kubernetes.io/projected/292f9a97-273c-431e-9c47-c0890d07c4ce-kube-api-access-nxq2j\") pod \"nova-metadata-0\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " pod="openstack/nova-metadata-0" Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.011816 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfzqw\" (UniqueName: \"kubernetes.io/projected/330175a1-2b61-4ad3-b170-9ee626141d7c-kube-api-access-rfzqw\") pod \"nova-scheduler-0\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.056080 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.152629 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.212525 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.631942 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.742025 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5","Type":"ContainerStarted","Data":"f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37"} Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.742106 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5","Type":"ContainerStarted","Data":"49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac"} Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.742121 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5","Type":"ContainerStarted","Data":"5f6ff3f191d7ae928cd5e52068e54be727c46c98917e914de6ff8caba817d11e"} Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.744773 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"292f9a97-273c-431e-9c47-c0890d07c4ce","Type":"ContainerStarted","Data":"e4c437bf9e2ae0f4be721f6fb80f8a575c0eff7df653889af5bfb919168c3ee6"} Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.764429 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.764405734 podStartE2EDuration="1.764405734s" podCreationTimestamp="2025-12-06 08:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:23:02.762104213 +0000 UTC m=+5396.216965257" watchObservedRunningTime="2025-12-06 08:23:02.764405734 +0000 UTC m=+5396.219266778" Dec 06 08:23:02 crc kubenswrapper[4945]: W1206 08:23:02.779836 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod330175a1_2b61_4ad3_b170_9ee626141d7c.slice/crio-d8cd65da57098ec58a484dd448a56e250dd965c9931c61ebad448b7871a0b921 WatchSource:0}: Error finding container d8cd65da57098ec58a484dd448a56e250dd965c9931c61ebad448b7871a0b921: Status 404 returned error can't find the container with id d8cd65da57098ec58a484dd448a56e250dd965c9931c61ebad448b7871a0b921 Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.782056 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.965452 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9542d8-84fc-4151-b31f-f63401d22f45" path="/var/lib/kubelet/pods/0c9542d8-84fc-4151-b31f-f63401d22f45/volumes" Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.966340 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc9d35a-31d2-4fbf-a605-08825bb7c0c5" path="/var/lib/kubelet/pods/1bc9d35a-31d2-4fbf-a605-08825bb7c0c5/volumes" Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.966956 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2679acb2-90dd-4a75-9e58-11b73546c8b7" path="/var/lib/kubelet/pods/2679acb2-90dd-4a75-9e58-11b73546c8b7/volumes" Dec 06 08:23:02 crc kubenswrapper[4945]: I1206 08:23:02.967989 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e137d224-e766-4063-bf00-85169d5adfa1" path="/var/lib/kubelet/pods/e137d224-e766-4063-bf00-85169d5adfa1/volumes" Dec 06 08:23:03 crc kubenswrapper[4945]: I1206 08:23:03.760737 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"330175a1-2b61-4ad3-b170-9ee626141d7c","Type":"ContainerStarted","Data":"46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b"} Dec 06 08:23:03 crc kubenswrapper[4945]: I1206 08:23:03.760781 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"330175a1-2b61-4ad3-b170-9ee626141d7c","Type":"ContainerStarted","Data":"d8cd65da57098ec58a484dd448a56e250dd965c9931c61ebad448b7871a0b921"} Dec 06 08:23:03 crc kubenswrapper[4945]: I1206 08:23:03.763326 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"292f9a97-273c-431e-9c47-c0890d07c4ce","Type":"ContainerStarted","Data":"038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337"} Dec 06 08:23:03 crc kubenswrapper[4945]: I1206 08:23:03.763382 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"292f9a97-273c-431e-9c47-c0890d07c4ce","Type":"ContainerStarted","Data":"2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124"} Dec 06 08:23:03 crc kubenswrapper[4945]: I1206 08:23:03.778713 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.778691961 podStartE2EDuration="2.778691961s" podCreationTimestamp="2025-12-06 08:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:23:03.774131689 +0000 UTC m=+5397.228992733" watchObservedRunningTime="2025-12-06 08:23:03.778691961 +0000 UTC m=+5397.233553005" Dec 06 08:23:03 crc kubenswrapper[4945]: I1206 08:23:03.802406 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.802387355 podStartE2EDuration="2.802387355s" podCreationTimestamp="2025-12-06 08:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:23:03.797788932 +0000 UTC m=+5397.252649996" watchObservedRunningTime="2025-12-06 08:23:03.802387355 +0000 UTC m=+5397.257248409" Dec 06 08:23:06 crc kubenswrapper[4945]: I1206 08:23:06.795879 4945 generic.go:334] "Generic (PLEG): container finished" podID="a133569e-416b-46a2-ab24-c1cfe9d4718d" containerID="d764b626224cc9f20ff1c6240e8386ae706bb83d4cef60d7c3aa322dc369c78e" exitCode=0 Dec 06 08:23:06 crc kubenswrapper[4945]: I1206 08:23:06.795953 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kj84k" event={"ID":"a133569e-416b-46a2-ab24-c1cfe9d4718d","Type":"ContainerDied","Data":"d764b626224cc9f20ff1c6240e8386ae706bb83d4cef60d7c3aa322dc369c78e"} Dec 06 08:23:07 crc kubenswrapper[4945]: I1206 08:23:07.153457 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 08:23:07 crc kubenswrapper[4945]: I1206 08:23:07.153509 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 08:23:07 crc kubenswrapper[4945]: I1206 08:23:07.212954 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.142420 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.233725 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-scripts\") pod \"a133569e-416b-46a2-ab24-c1cfe9d4718d\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.233824 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-combined-ca-bundle\") pod \"a133569e-416b-46a2-ab24-c1cfe9d4718d\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.234045 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-config-data\") pod \"a133569e-416b-46a2-ab24-c1cfe9d4718d\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.234140 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb2jn\" (UniqueName: \"kubernetes.io/projected/a133569e-416b-46a2-ab24-c1cfe9d4718d-kube-api-access-zb2jn\") pod \"a133569e-416b-46a2-ab24-c1cfe9d4718d\" (UID: \"a133569e-416b-46a2-ab24-c1cfe9d4718d\") " Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.239647 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-scripts" (OuterVolumeSpecName: "scripts") pod "a133569e-416b-46a2-ab24-c1cfe9d4718d" (UID: "a133569e-416b-46a2-ab24-c1cfe9d4718d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.247580 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a133569e-416b-46a2-ab24-c1cfe9d4718d-kube-api-access-zb2jn" (OuterVolumeSpecName: "kube-api-access-zb2jn") pod "a133569e-416b-46a2-ab24-c1cfe9d4718d" (UID: "a133569e-416b-46a2-ab24-c1cfe9d4718d"). InnerVolumeSpecName "kube-api-access-zb2jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.261214 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-config-data" (OuterVolumeSpecName: "config-data") pod "a133569e-416b-46a2-ab24-c1cfe9d4718d" (UID: "a133569e-416b-46a2-ab24-c1cfe9d4718d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.261488 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a133569e-416b-46a2-ab24-c1cfe9d4718d" (UID: "a133569e-416b-46a2-ab24-c1cfe9d4718d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.336773 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.336813 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.336825 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a133569e-416b-46a2-ab24-c1cfe9d4718d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.336837 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb2jn\" (UniqueName: \"kubernetes.io/projected/a133569e-416b-46a2-ab24-c1cfe9d4718d-kube-api-access-zb2jn\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.814340 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kj84k" event={"ID":"a133569e-416b-46a2-ab24-c1cfe9d4718d","Type":"ContainerDied","Data":"be7f99edc2056b20f8850f1d8c677e6a2ec52613ee423798f925702c6016c5c4"} Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.814732 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be7f99edc2056b20f8850f1d8c677e6a2ec52613ee423798f925702c6016c5c4" Dec 06 08:23:08 crc kubenswrapper[4945]: I1206 08:23:08.814491 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kj84k" Dec 06 08:23:09 crc kubenswrapper[4945]: I1206 08:23:09.001490 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:09 crc kubenswrapper[4945]: I1206 08:23:09.001825 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" containerName="nova-api-log" containerID="cri-o://49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac" gracePeriod=30 Dec 06 08:23:09 crc kubenswrapper[4945]: I1206 08:23:09.002178 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" containerName="nova-api-api" containerID="cri-o://f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37" gracePeriod=30 Dec 06 08:23:09 crc kubenswrapper[4945]: I1206 08:23:09.080372 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:09 crc kubenswrapper[4945]: I1206 08:23:09.080631 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="330175a1-2b61-4ad3-b170-9ee626141d7c" containerName="nova-scheduler-scheduler" containerID="cri-o://46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b" gracePeriod=30 Dec 06 08:23:09 crc kubenswrapper[4945]: I1206 08:23:09.094620 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:09 crc kubenswrapper[4945]: I1206 08:23:09.095009 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="292f9a97-273c-431e-9c47-c0890d07c4ce" containerName="nova-metadata-log" containerID="cri-o://2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124" gracePeriod=30 Dec 06 08:23:09 crc kubenswrapper[4945]: I1206 08:23:09.095106 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="292f9a97-273c-431e-9c47-c0890d07c4ce" containerName="nova-metadata-metadata" containerID="cri-o://038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337" gracePeriod=30 Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.524599 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.562386 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-logs\") pod \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.562643 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-combined-ca-bundle\") pod \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.562758 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4kfz\" (UniqueName: \"kubernetes.io/projected/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-kube-api-access-k4kfz\") pod \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.562780 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-logs" (OuterVolumeSpecName: "logs") pod "fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" (UID: "fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.562873 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-config-data\") pod \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\" (UID: \"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.563748 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.567636 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-kube-api-access-k4kfz" (OuterVolumeSpecName: "kube-api-access-k4kfz") pod "fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" (UID: "fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5"). InnerVolumeSpecName "kube-api-access-k4kfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.588750 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-config-data" (OuterVolumeSpecName: "config-data") pod "fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" (UID: "fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.594647 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.598601 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" (UID: "fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.665213 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-combined-ca-bundle\") pod \"292f9a97-273c-431e-9c47-c0890d07c4ce\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.665350 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292f9a97-273c-431e-9c47-c0890d07c4ce-logs\") pod \"292f9a97-273c-431e-9c47-c0890d07c4ce\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.665417 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-config-data\") pod \"292f9a97-273c-431e-9c47-c0890d07c4ce\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.665531 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxq2j\" (UniqueName: \"kubernetes.io/projected/292f9a97-273c-431e-9c47-c0890d07c4ce-kube-api-access-nxq2j\") pod \"292f9a97-273c-431e-9c47-c0890d07c4ce\" (UID: \"292f9a97-273c-431e-9c47-c0890d07c4ce\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.665911 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.665926 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4kfz\" (UniqueName: \"kubernetes.io/projected/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-kube-api-access-k4kfz\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.665940 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.666673 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/292f9a97-273c-431e-9c47-c0890d07c4ce-logs" (OuterVolumeSpecName: "logs") pod "292f9a97-273c-431e-9c47-c0890d07c4ce" (UID: "292f9a97-273c-431e-9c47-c0890d07c4ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.669719 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/292f9a97-273c-431e-9c47-c0890d07c4ce-kube-api-access-nxq2j" (OuterVolumeSpecName: "kube-api-access-nxq2j") pod "292f9a97-273c-431e-9c47-c0890d07c4ce" (UID: "292f9a97-273c-431e-9c47-c0890d07c4ce"). InnerVolumeSpecName "kube-api-access-nxq2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.701971 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-config-data" (OuterVolumeSpecName: "config-data") pod "292f9a97-273c-431e-9c47-c0890d07c4ce" (UID: "292f9a97-273c-431e-9c47-c0890d07c4ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.725484 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "292f9a97-273c-431e-9c47-c0890d07c4ce" (UID: "292f9a97-273c-431e-9c47-c0890d07c4ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.767591 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxq2j\" (UniqueName: \"kubernetes.io/projected/292f9a97-273c-431e-9c47-c0890d07c4ce-kube-api-access-nxq2j\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.767640 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.767655 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292f9a97-273c-431e-9c47-c0890d07c4ce-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.767666 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292f9a97-273c-431e-9c47-c0890d07c4ce-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.824887 4945 generic.go:334] "Generic (PLEG): container finished" podID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" containerID="f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37" exitCode=0 Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.824919 4945 generic.go:334] "Generic (PLEG): container finished" podID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" containerID="49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac" exitCode=143 Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.824967 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5","Type":"ContainerDied","Data":"f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37"} Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.824999 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5","Type":"ContainerDied","Data":"49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac"} Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.825014 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5","Type":"ContainerDied","Data":"5f6ff3f191d7ae928cd5e52068e54be727c46c98917e914de6ff8caba817d11e"} Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.825031 4945 scope.go:117] "RemoveContainer" containerID="f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.825179 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.827667 4945 generic.go:334] "Generic (PLEG): container finished" podID="292f9a97-273c-431e-9c47-c0890d07c4ce" containerID="038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337" exitCode=0 Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.827693 4945 generic.go:334] "Generic (PLEG): container finished" podID="292f9a97-273c-431e-9c47-c0890d07c4ce" containerID="2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124" exitCode=143 Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.827715 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"292f9a97-273c-431e-9c47-c0890d07c4ce","Type":"ContainerDied","Data":"038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337"} Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.827725 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.827741 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"292f9a97-273c-431e-9c47-c0890d07c4ce","Type":"ContainerDied","Data":"2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124"} Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.827752 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"292f9a97-273c-431e-9c47-c0890d07c4ce","Type":"ContainerDied","Data":"e4c437bf9e2ae0f4be721f6fb80f8a575c0eff7df653889af5bfb919168c3ee6"} Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.861560 4945 scope.go:117] "RemoveContainer" containerID="49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.864533 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.878375 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.888726 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898146 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:09.898605 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" containerName="nova-api-api" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898617 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" containerName="nova-api-api" Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:09.898625 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292f9a97-273c-431e-9c47-c0890d07c4ce" containerName="nova-metadata-log" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898631 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="292f9a97-273c-431e-9c47-c0890d07c4ce" containerName="nova-metadata-log" Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:09.898656 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" containerName="nova-api-log" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898663 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" containerName="nova-api-log" Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:09.898672 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a133569e-416b-46a2-ab24-c1cfe9d4718d" containerName="nova-manage" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898678 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a133569e-416b-46a2-ab24-c1cfe9d4718d" containerName="nova-manage" Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:09.898699 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292f9a97-273c-431e-9c47-c0890d07c4ce" containerName="nova-metadata-metadata" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898704 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="292f9a97-273c-431e-9c47-c0890d07c4ce" containerName="nova-metadata-metadata" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898868 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="292f9a97-273c-431e-9c47-c0890d07c4ce" containerName="nova-metadata-metadata" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898881 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a133569e-416b-46a2-ab24-c1cfe9d4718d" containerName="nova-manage" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898891 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" containerName="nova-api-api" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898912 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" containerName="nova-api-log" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.898920 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="292f9a97-273c-431e-9c47-c0890d07c4ce" containerName="nova-metadata-log" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.899908 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.903343 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.916218 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.925588 4945 scope.go:117] "RemoveContainer" containerID="f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37" Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:09.926504 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37\": container with ID starting with f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37 not found: ID does not exist" containerID="f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.926546 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37"} err="failed to get container status \"f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37\": rpc error: code = NotFound desc = could not find container \"f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37\": container with ID starting with f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37 not found: ID does not exist" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.926575 4945 scope.go:117] "RemoveContainer" containerID="49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac" Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:09.926809 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac\": container with ID starting with 49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac not found: ID does not exist" containerID="49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.926844 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac"} err="failed to get container status \"49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac\": rpc error: code = NotFound desc = could not find container \"49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac\": container with ID starting with 49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac not found: ID does not exist" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.926861 4945 scope.go:117] "RemoveContainer" containerID="f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.927052 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37"} err="failed to get container status \"f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37\": rpc error: code = NotFound desc = could not find container \"f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37\": container with ID starting with f8722911690d9c200e7c4b3825c864b7f9f07c515c44b19c02b7905a8031cc37 not found: ID does not exist" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.927070 4945 scope.go:117] "RemoveContainer" containerID="49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.927267 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac"} err="failed to get container status \"49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac\": rpc error: code = NotFound desc = could not find container \"49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac\": container with ID starting with 49f405d1572dec0dbb3f752e34f3bb94a5ea2f2799224a33f615945b5d869aac not found: ID does not exist" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.927331 4945 scope.go:117] "RemoveContainer" containerID="038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.937803 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.951368 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.952929 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.953534 4945 scope.go:117] "RemoveContainer" containerID="2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.955701 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.965363 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.974885 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-config-data\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.974941 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.975190 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-config-data\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.975243 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-logs\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.975308 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgsfm\" (UniqueName: \"kubernetes.io/projected/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-kube-api-access-vgsfm\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.975415 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8475348b-c57c-439a-ac05-1a71be4b29b8-logs\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.975484 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.975539 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cftd\" (UniqueName: \"kubernetes.io/projected/8475348b-c57c-439a-ac05-1a71be4b29b8-kube-api-access-6cftd\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.981389 4945 scope.go:117] "RemoveContainer" containerID="038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337" Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:09.981971 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337\": container with ID starting with 038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337 not found: ID does not exist" containerID="038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.982002 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337"} err="failed to get container status \"038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337\": rpc error: code = NotFound desc = could not find container \"038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337\": container with ID starting with 038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337 not found: ID does not exist" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.982021 4945 scope.go:117] "RemoveContainer" containerID="2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124" Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:09.982326 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124\": container with ID starting with 2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124 not found: ID does not exist" containerID="2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.982363 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124"} err="failed to get container status \"2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124\": rpc error: code = NotFound desc = could not find container \"2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124\": container with ID starting with 2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124 not found: ID does not exist" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.982390 4945 scope.go:117] "RemoveContainer" containerID="038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.982676 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337"} err="failed to get container status \"038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337\": rpc error: code = NotFound desc = could not find container \"038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337\": container with ID starting with 038baa64aed42914eee002477d1ea6494ee983f461d1ac366fb258c516fd3337 not found: ID does not exist" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.982692 4945 scope.go:117] "RemoveContainer" containerID="2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:09.982914 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124"} err="failed to get container status \"2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124\": rpc error: code = NotFound desc = could not find container \"2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124\": container with ID starting with 2ef68afbfe47e998b210526a8860f5596d61e1b363f43d3c1e54243b9e204124 not found: ID does not exist" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.077798 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.077869 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cftd\" (UniqueName: \"kubernetes.io/projected/8475348b-c57c-439a-ac05-1a71be4b29b8-kube-api-access-6cftd\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.077941 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-config-data\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.077977 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.078060 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-config-data\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.078088 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-logs\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.078119 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgsfm\" (UniqueName: \"kubernetes.io/projected/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-kube-api-access-vgsfm\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.078176 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8475348b-c57c-439a-ac05-1a71be4b29b8-logs\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.078546 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8475348b-c57c-439a-ac05-1a71be4b29b8-logs\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.078935 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-logs\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.083398 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.084684 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-config-data\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.085313 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.086422 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-config-data\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.095968 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cftd\" (UniqueName: \"kubernetes.io/projected/8475348b-c57c-439a-ac05-1a71be4b29b8-kube-api-access-6cftd\") pod \"nova-api-0\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.096865 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgsfm\" (UniqueName: \"kubernetes.io/projected/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-kube-api-access-vgsfm\") pod \"nova-metadata-0\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.244483 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.361422 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.529266 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.585421 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfzqw\" (UniqueName: \"kubernetes.io/projected/330175a1-2b61-4ad3-b170-9ee626141d7c-kube-api-access-rfzqw\") pod \"330175a1-2b61-4ad3-b170-9ee626141d7c\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.587697 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-config-data\") pod \"330175a1-2b61-4ad3-b170-9ee626141d7c\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.587778 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-combined-ca-bundle\") pod \"330175a1-2b61-4ad3-b170-9ee626141d7c\" (UID: \"330175a1-2b61-4ad3-b170-9ee626141d7c\") " Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.592293 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/330175a1-2b61-4ad3-b170-9ee626141d7c-kube-api-access-rfzqw" (OuterVolumeSpecName: "kube-api-access-rfzqw") pod "330175a1-2b61-4ad3-b170-9ee626141d7c" (UID: "330175a1-2b61-4ad3-b170-9ee626141d7c"). InnerVolumeSpecName "kube-api-access-rfzqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.610689 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-config-data" (OuterVolumeSpecName: "config-data") pod "330175a1-2b61-4ad3-b170-9ee626141d7c" (UID: "330175a1-2b61-4ad3-b170-9ee626141d7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.619150 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "330175a1-2b61-4ad3-b170-9ee626141d7c" (UID: "330175a1-2b61-4ad3-b170-9ee626141d7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.690089 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.690126 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/330175a1-2b61-4ad3-b170-9ee626141d7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.690138 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfzqw\" (UniqueName: \"kubernetes.io/projected/330175a1-2b61-4ad3-b170-9ee626141d7c-kube-api-access-rfzqw\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.711815 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: W1206 08:23:10.715611 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bb5ebc7_fa62_420c_b3a3_ff65df244f71.slice/crio-8cafdaae2eab9092135b9c4d9983fe2794d79f77866a7dd94b4ab60df6a3f7fb WatchSource:0}: Error finding container 8cafdaae2eab9092135b9c4d9983fe2794d79f77866a7dd94b4ab60df6a3f7fb: Status 404 returned error can't find the container with id 8cafdaae2eab9092135b9c4d9983fe2794d79f77866a7dd94b4ab60df6a3f7fb Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.842498 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.846828 4945 generic.go:334] "Generic (PLEG): container finished" podID="330175a1-2b61-4ad3-b170-9ee626141d7c" containerID="46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b" exitCode=0 Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.846884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"330175a1-2b61-4ad3-b170-9ee626141d7c","Type":"ContainerDied","Data":"46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b"} Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.846894 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.846909 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"330175a1-2b61-4ad3-b170-9ee626141d7c","Type":"ContainerDied","Data":"d8cd65da57098ec58a484dd448a56e250dd965c9931c61ebad448b7871a0b921"} Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.846924 4945 scope.go:117] "RemoveContainer" containerID="46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b" Dec 06 08:23:10 crc kubenswrapper[4945]: W1206 08:23:10.847952 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8475348b_c57c_439a_ac05_1a71be4b29b8.slice/crio-1a56a562e47c14be1a7655a9274f333dc353bdc22be42a79f993a20f7f87dfe2 WatchSource:0}: Error finding container 1a56a562e47c14be1a7655a9274f333dc353bdc22be42a79f993a20f7f87dfe2: Status 404 returned error can't find the container with id 1a56a562e47c14be1a7655a9274f333dc353bdc22be42a79f993a20f7f87dfe2 Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.854869 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6bb5ebc7-fa62-420c-b3a3-ff65df244f71","Type":"ContainerStarted","Data":"8cafdaae2eab9092135b9c4d9983fe2794d79f77866a7dd94b4ab60df6a3f7fb"} Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.878556 4945 scope.go:117] "RemoveContainer" containerID="46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b" Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:10.879067 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b\": container with ID starting with 46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b not found: ID does not exist" containerID="46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.879105 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b"} err="failed to get container status \"46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b\": rpc error: code = NotFound desc = could not find container \"46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b\": container with ID starting with 46b4db051f4d43f8a975bab6c058e21a5ae8c21ced5b859e135b545b81d4728b not found: ID does not exist" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.879483 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.893858 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.904807 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: E1206 08:23:10.905279 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330175a1-2b61-4ad3-b170-9ee626141d7c" containerName="nova-scheduler-scheduler" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.905385 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="330175a1-2b61-4ad3-b170-9ee626141d7c" containerName="nova-scheduler-scheduler" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.906225 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="330175a1-2b61-4ad3-b170-9ee626141d7c" containerName="nova-scheduler-scheduler" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.906967 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.909080 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.932165 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.965008 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="292f9a97-273c-431e-9c47-c0890d07c4ce" path="/var/lib/kubelet/pods/292f9a97-273c-431e-9c47-c0890d07c4ce/volumes" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.965589 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="330175a1-2b61-4ad3-b170-9ee626141d7c" path="/var/lib/kubelet/pods/330175a1-2b61-4ad3-b170-9ee626141d7c/volumes" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.966258 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5" path="/var/lib/kubelet/pods/fbb20fdb-3c06-44ad-837f-3eb8d32bbdb5/volumes" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.996207 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.996266 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-config-data\") pod \"nova-scheduler-0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:10 crc kubenswrapper[4945]: I1206 08:23:10.996321 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpj2h\" (UniqueName: \"kubernetes.io/projected/299c3108-fdd9-413b-a0cc-b751515e0ad0-kube-api-access-kpj2h\") pod \"nova-scheduler-0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.097421 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.097461 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-config-data\") pod \"nova-scheduler-0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.097484 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpj2h\" (UniqueName: \"kubernetes.io/projected/299c3108-fdd9-413b-a0cc-b751515e0ad0-kube-api-access-kpj2h\") pod \"nova-scheduler-0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.102150 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.102507 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-config-data\") pod \"nova-scheduler-0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.113144 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpj2h\" (UniqueName: \"kubernetes.io/projected/299c3108-fdd9-413b-a0cc-b751515e0ad0-kube-api-access-kpj2h\") pod \"nova-scheduler-0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " pod="openstack/nova-scheduler-0" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.228870 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.650235 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:23:11 crc kubenswrapper[4945]: W1206 08:23:11.653297 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod299c3108_fdd9_413b_a0cc_b751515e0ad0.slice/crio-e0cad7e62909687eee6c985804ff5f031778fd1e23f1a55848f48880b96c9432 WatchSource:0}: Error finding container e0cad7e62909687eee6c985804ff5f031778fd1e23f1a55848f48880b96c9432: Status 404 returned error can't find the container with id e0cad7e62909687eee6c985804ff5f031778fd1e23f1a55848f48880b96c9432 Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.871078 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6bb5ebc7-fa62-420c-b3a3-ff65df244f71","Type":"ContainerStarted","Data":"530de7d699d2067771b77ed69a84ff82d666a14f233167c207ba9271e9f2f8e5"} Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.871131 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6bb5ebc7-fa62-420c-b3a3-ff65df244f71","Type":"ContainerStarted","Data":"37fb7843aa59c006c625d448b556112a9a74ef369c01edcabd5789b8b7039f97"} Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.877708 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8475348b-c57c-439a-ac05-1a71be4b29b8","Type":"ContainerStarted","Data":"1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145"} Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.877752 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8475348b-c57c-439a-ac05-1a71be4b29b8","Type":"ContainerStarted","Data":"5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff"} Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.877766 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8475348b-c57c-439a-ac05-1a71be4b29b8","Type":"ContainerStarted","Data":"1a56a562e47c14be1a7655a9274f333dc353bdc22be42a79f993a20f7f87dfe2"} Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.879697 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"299c3108-fdd9-413b-a0cc-b751515e0ad0","Type":"ContainerStarted","Data":"a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d"} Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.879721 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"299c3108-fdd9-413b-a0cc-b751515e0ad0","Type":"ContainerStarted","Data":"e0cad7e62909687eee6c985804ff5f031778fd1e23f1a55848f48880b96c9432"} Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.890811 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.890791188 podStartE2EDuration="2.890791188s" podCreationTimestamp="2025-12-06 08:23:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:23:11.888091335 +0000 UTC m=+5405.342952379" watchObservedRunningTime="2025-12-06 08:23:11.890791188 +0000 UTC m=+5405.345652222" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.913142 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.913122895 podStartE2EDuration="2.913122895s" podCreationTimestamp="2025-12-06 08:23:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:23:11.912598531 +0000 UTC m=+5405.367459605" watchObservedRunningTime="2025-12-06 08:23:11.913122895 +0000 UTC m=+5405.367983929" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.931791 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.931770984 podStartE2EDuration="1.931770984s" podCreationTimestamp="2025-12-06 08:23:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:23:11.928404364 +0000 UTC m=+5405.383265428" watchObservedRunningTime="2025-12-06 08:23:11.931770984 +0000 UTC m=+5405.386632028" Dec 06 08:23:11 crc kubenswrapper[4945]: I1206 08:23:11.955765 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:23:11 crc kubenswrapper[4945]: E1206 08:23:11.956085 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:23:15 crc kubenswrapper[4945]: I1206 08:23:15.244658 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 08:23:15 crc kubenswrapper[4945]: I1206 08:23:15.245096 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 08:23:16 crc kubenswrapper[4945]: I1206 08:23:16.229018 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 08:23:20 crc kubenswrapper[4945]: I1206 08:23:20.245553 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 08:23:20 crc kubenswrapper[4945]: I1206 08:23:20.247556 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 08:23:20 crc kubenswrapper[4945]: I1206 08:23:20.363074 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 08:23:20 crc kubenswrapper[4945]: I1206 08:23:20.363133 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 08:23:21 crc kubenswrapper[4945]: I1206 08:23:21.229688 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 08:23:21 crc kubenswrapper[4945]: I1206 08:23:21.262534 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 08:23:21 crc kubenswrapper[4945]: I1206 08:23:21.286533 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:23:21 crc kubenswrapper[4945]: I1206 08:23:21.328552 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:23:21 crc kubenswrapper[4945]: I1206 08:23:21.446465 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.70:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:23:21 crc kubenswrapper[4945]: I1206 08:23:21.446721 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.70:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:23:22 crc kubenswrapper[4945]: I1206 08:23:22.013534 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 08:23:26 crc kubenswrapper[4945]: I1206 08:23:26.963032 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:23:26 crc kubenswrapper[4945]: E1206 08:23:26.963942 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:23:30 crc kubenswrapper[4945]: I1206 08:23:30.247692 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 08:23:30 crc kubenswrapper[4945]: I1206 08:23:30.248293 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 08:23:30 crc kubenswrapper[4945]: I1206 08:23:30.250656 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 08:23:30 crc kubenswrapper[4945]: I1206 08:23:30.250897 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 08:23:30 crc kubenswrapper[4945]: I1206 08:23:30.375811 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 08:23:30 crc kubenswrapper[4945]: I1206 08:23:30.376798 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 08:23:30 crc kubenswrapper[4945]: I1206 08:23:30.379349 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 08:23:30 crc kubenswrapper[4945]: I1206 08:23:30.379608 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.093647 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.097020 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.277303 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98f9d44d7-vnm9b"] Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.280893 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.293515 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98f9d44d7-vnm9b"] Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.437395 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-sb\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.437509 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-config\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.437564 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-dns-svc\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.437589 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbvc7\" (UniqueName: \"kubernetes.io/projected/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-kube-api-access-gbvc7\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.437709 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-nb\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.539442 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-nb\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.539514 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-sb\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.539579 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-config\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.539600 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-dns-svc\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.539619 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbvc7\" (UniqueName: \"kubernetes.io/projected/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-kube-api-access-gbvc7\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.540705 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-dns-svc\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.540772 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-config\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.540776 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-nb\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.541254 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-sb\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.562677 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbvc7\" (UniqueName: \"kubernetes.io/projected/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-kube-api-access-gbvc7\") pod \"dnsmasq-dns-98f9d44d7-vnm9b\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:31 crc kubenswrapper[4945]: I1206 08:23:31.605070 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:32 crc kubenswrapper[4945]: I1206 08:23:32.081028 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98f9d44d7-vnm9b"] Dec 06 08:23:32 crc kubenswrapper[4945]: W1206 08:23:32.085466 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0a7f6b6_8b5c_42b6_b813_57fd600fa94c.slice/crio-c5351124c621936b558b184a8e6a7be180bb3b02a7899f370f8dabbf0a070966 WatchSource:0}: Error finding container c5351124c621936b558b184a8e6a7be180bb3b02a7899f370f8dabbf0a070966: Status 404 returned error can't find the container with id c5351124c621936b558b184a8e6a7be180bb3b02a7899f370f8dabbf0a070966 Dec 06 08:23:32 crc kubenswrapper[4945]: I1206 08:23:32.112475 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" event={"ID":"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c","Type":"ContainerStarted","Data":"c5351124c621936b558b184a8e6a7be180bb3b02a7899f370f8dabbf0a070966"} Dec 06 08:23:33 crc kubenswrapper[4945]: I1206 08:23:33.121822 4945 generic.go:334] "Generic (PLEG): container finished" podID="b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" containerID="531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904" exitCode=0 Dec 06 08:23:33 crc kubenswrapper[4945]: I1206 08:23:33.121925 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" event={"ID":"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c","Type":"ContainerDied","Data":"531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904"} Dec 06 08:23:34 crc kubenswrapper[4945]: I1206 08:23:34.133105 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" event={"ID":"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c","Type":"ContainerStarted","Data":"00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80"} Dec 06 08:23:34 crc kubenswrapper[4945]: I1206 08:23:34.133733 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:34 crc kubenswrapper[4945]: I1206 08:23:34.159931 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" podStartSLOduration=3.159907375 podStartE2EDuration="3.159907375s" podCreationTimestamp="2025-12-06 08:23:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:23:34.151971802 +0000 UTC m=+5427.606832846" watchObservedRunningTime="2025-12-06 08:23:34.159907375 +0000 UTC m=+5427.614768419" Dec 06 08:23:39 crc kubenswrapper[4945]: I1206 08:23:39.953791 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:23:39 crc kubenswrapper[4945]: E1206 08:23:39.954670 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:23:41 crc kubenswrapper[4945]: I1206 08:23:41.606561 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:23:41 crc kubenswrapper[4945]: I1206 08:23:41.704389 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7879cc8c79-sn2f4"] Dec 06 08:23:41 crc kubenswrapper[4945]: I1206 08:23:41.704877 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" podUID="72bcfa97-b158-4da4-ae32-8a24fba0062e" containerName="dnsmasq-dns" containerID="cri-o://e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c" gracePeriod=10 Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.193908 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.234193 4945 generic.go:334] "Generic (PLEG): container finished" podID="72bcfa97-b158-4da4-ae32-8a24fba0062e" containerID="e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c" exitCode=0 Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.234235 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" event={"ID":"72bcfa97-b158-4da4-ae32-8a24fba0062e","Type":"ContainerDied","Data":"e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c"} Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.234262 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" event={"ID":"72bcfa97-b158-4da4-ae32-8a24fba0062e","Type":"ContainerDied","Data":"23dadb8a4f5580ffbeae633266ae4088e885b0dc60d3b642c5197848635de704"} Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.234296 4945 scope.go:117] "RemoveContainer" containerID="e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.234441 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7879cc8c79-sn2f4" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.253993 4945 scope.go:117] "RemoveContainer" containerID="ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.272302 4945 scope.go:117] "RemoveContainer" containerID="e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c" Dec 06 08:23:42 crc kubenswrapper[4945]: E1206 08:23:42.272723 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c\": container with ID starting with e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c not found: ID does not exist" containerID="e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.272785 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c"} err="failed to get container status \"e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c\": rpc error: code = NotFound desc = could not find container \"e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c\": container with ID starting with e7d1dd317fc2f60900a7c7191ecdae6b7ec37ed5d4c6412ef935890c09c7725c not found: ID does not exist" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.272821 4945 scope.go:117] "RemoveContainer" containerID="ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49" Dec 06 08:23:42 crc kubenswrapper[4945]: E1206 08:23:42.273189 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49\": container with ID starting with ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49 not found: ID does not exist" containerID="ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.273229 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49"} err="failed to get container status \"ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49\": rpc error: code = NotFound desc = could not find container \"ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49\": container with ID starting with ed7a0925815d0c996afa58e3f2a331e6162aaaff028a13ed1d9c38f7bb39ce49 not found: ID does not exist" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.355987 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-nb\") pod \"72bcfa97-b158-4da4-ae32-8a24fba0062e\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.356050 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-sb\") pod \"72bcfa97-b158-4da4-ae32-8a24fba0062e\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.356259 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vqdd\" (UniqueName: \"kubernetes.io/projected/72bcfa97-b158-4da4-ae32-8a24fba0062e-kube-api-access-5vqdd\") pod \"72bcfa97-b158-4da4-ae32-8a24fba0062e\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.356319 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-config\") pod \"72bcfa97-b158-4da4-ae32-8a24fba0062e\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.356345 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-dns-svc\") pod \"72bcfa97-b158-4da4-ae32-8a24fba0062e\" (UID: \"72bcfa97-b158-4da4-ae32-8a24fba0062e\") " Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.372587 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72bcfa97-b158-4da4-ae32-8a24fba0062e-kube-api-access-5vqdd" (OuterVolumeSpecName: "kube-api-access-5vqdd") pod "72bcfa97-b158-4da4-ae32-8a24fba0062e" (UID: "72bcfa97-b158-4da4-ae32-8a24fba0062e"). InnerVolumeSpecName "kube-api-access-5vqdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.401062 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "72bcfa97-b158-4da4-ae32-8a24fba0062e" (UID: "72bcfa97-b158-4da4-ae32-8a24fba0062e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.406460 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "72bcfa97-b158-4da4-ae32-8a24fba0062e" (UID: "72bcfa97-b158-4da4-ae32-8a24fba0062e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.412014 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-config" (OuterVolumeSpecName: "config") pod "72bcfa97-b158-4da4-ae32-8a24fba0062e" (UID: "72bcfa97-b158-4da4-ae32-8a24fba0062e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.412758 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "72bcfa97-b158-4da4-ae32-8a24fba0062e" (UID: "72bcfa97-b158-4da4-ae32-8a24fba0062e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.458259 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vqdd\" (UniqueName: \"kubernetes.io/projected/72bcfa97-b158-4da4-ae32-8a24fba0062e-kube-api-access-5vqdd\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.458330 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.458346 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.458364 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.458381 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72bcfa97-b158-4da4-ae32-8a24fba0062e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.575134 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7879cc8c79-sn2f4"] Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.609506 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7879cc8c79-sn2f4"] Dec 06 08:23:42 crc kubenswrapper[4945]: I1206 08:23:42.964194 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72bcfa97-b158-4da4-ae32-8a24fba0062e" path="/var/lib/kubelet/pods/72bcfa97-b158-4da4-ae32-8a24fba0062e/volumes" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.003581 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-fmsp8"] Dec 06 08:23:44 crc kubenswrapper[4945]: E1206 08:23:44.004212 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72bcfa97-b158-4da4-ae32-8a24fba0062e" containerName="init" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.004224 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bcfa97-b158-4da4-ae32-8a24fba0062e" containerName="init" Dec 06 08:23:44 crc kubenswrapper[4945]: E1206 08:23:44.004253 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72bcfa97-b158-4da4-ae32-8a24fba0062e" containerName="dnsmasq-dns" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.004259 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bcfa97-b158-4da4-ae32-8a24fba0062e" containerName="dnsmasq-dns" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.004465 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72bcfa97-b158-4da4-ae32-8a24fba0062e" containerName="dnsmasq-dns" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.005074 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-fmsp8" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.015072 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-fmsp8"] Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.110301 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-1115-account-create-update-qkllx"] Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.111828 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1115-account-create-update-qkllx" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.114739 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.120808 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1115-account-create-update-qkllx"] Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.193571 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-operator-scripts\") pod \"cinder-db-create-fmsp8\" (UID: \"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98\") " pod="openstack/cinder-db-create-fmsp8" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.193968 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgsrw\" (UniqueName: \"kubernetes.io/projected/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-kube-api-access-kgsrw\") pod \"cinder-db-create-fmsp8\" (UID: \"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98\") " pod="openstack/cinder-db-create-fmsp8" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.295694 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-operator-scripts\") pod \"cinder-db-create-fmsp8\" (UID: \"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98\") " pod="openstack/cinder-db-create-fmsp8" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.295962 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64n8q\" (UniqueName: \"kubernetes.io/projected/211109fe-4865-4082-ae88-8b4fb6facefa-kube-api-access-64n8q\") pod \"cinder-1115-account-create-update-qkllx\" (UID: \"211109fe-4865-4082-ae88-8b4fb6facefa\") " pod="openstack/cinder-1115-account-create-update-qkllx" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.296052 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgsrw\" (UniqueName: \"kubernetes.io/projected/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-kube-api-access-kgsrw\") pod \"cinder-db-create-fmsp8\" (UID: \"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98\") " pod="openstack/cinder-db-create-fmsp8" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.296148 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/211109fe-4865-4082-ae88-8b4fb6facefa-operator-scripts\") pod \"cinder-1115-account-create-update-qkllx\" (UID: \"211109fe-4865-4082-ae88-8b4fb6facefa\") " pod="openstack/cinder-1115-account-create-update-qkllx" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.296645 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-operator-scripts\") pod \"cinder-db-create-fmsp8\" (UID: \"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98\") " pod="openstack/cinder-db-create-fmsp8" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.316806 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgsrw\" (UniqueName: \"kubernetes.io/projected/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-kube-api-access-kgsrw\") pod \"cinder-db-create-fmsp8\" (UID: \"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98\") " pod="openstack/cinder-db-create-fmsp8" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.324806 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-fmsp8" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.398321 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64n8q\" (UniqueName: \"kubernetes.io/projected/211109fe-4865-4082-ae88-8b4fb6facefa-kube-api-access-64n8q\") pod \"cinder-1115-account-create-update-qkllx\" (UID: \"211109fe-4865-4082-ae88-8b4fb6facefa\") " pod="openstack/cinder-1115-account-create-update-qkllx" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.398401 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/211109fe-4865-4082-ae88-8b4fb6facefa-operator-scripts\") pod \"cinder-1115-account-create-update-qkllx\" (UID: \"211109fe-4865-4082-ae88-8b4fb6facefa\") " pod="openstack/cinder-1115-account-create-update-qkllx" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.399128 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/211109fe-4865-4082-ae88-8b4fb6facefa-operator-scripts\") pod \"cinder-1115-account-create-update-qkllx\" (UID: \"211109fe-4865-4082-ae88-8b4fb6facefa\") " pod="openstack/cinder-1115-account-create-update-qkllx" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.417315 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64n8q\" (UniqueName: \"kubernetes.io/projected/211109fe-4865-4082-ae88-8b4fb6facefa-kube-api-access-64n8q\") pod \"cinder-1115-account-create-update-qkllx\" (UID: \"211109fe-4865-4082-ae88-8b4fb6facefa\") " pod="openstack/cinder-1115-account-create-update-qkllx" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.427222 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1115-account-create-update-qkllx" Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.845766 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-fmsp8"] Dec 06 08:23:44 crc kubenswrapper[4945]: W1206 08:23:44.850588 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c4f1b77_f8f3_490a_bc9f_f7c6829a7c98.slice/crio-3a9a028740ae5eda20842be8f1ec8e497719f1e26f9b2cbc51204ca7f888bb59 WatchSource:0}: Error finding container 3a9a028740ae5eda20842be8f1ec8e497719f1e26f9b2cbc51204ca7f888bb59: Status 404 returned error can't find the container with id 3a9a028740ae5eda20842be8f1ec8e497719f1e26f9b2cbc51204ca7f888bb59 Dec 06 08:23:44 crc kubenswrapper[4945]: I1206 08:23:44.879814 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1115-account-create-update-qkllx"] Dec 06 08:23:44 crc kubenswrapper[4945]: W1206 08:23:44.886482 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod211109fe_4865_4082_ae88_8b4fb6facefa.slice/crio-fdfebbf4567440701848bb0f8067d0f66bf9b2305fb9307c25d6526e65fef417 WatchSource:0}: Error finding container fdfebbf4567440701848bb0f8067d0f66bf9b2305fb9307c25d6526e65fef417: Status 404 returned error can't find the container with id fdfebbf4567440701848bb0f8067d0f66bf9b2305fb9307c25d6526e65fef417 Dec 06 08:23:45 crc kubenswrapper[4945]: I1206 08:23:45.260820 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1115-account-create-update-qkllx" event={"ID":"211109fe-4865-4082-ae88-8b4fb6facefa","Type":"ContainerStarted","Data":"d1584e94d34ca3d0c00f1f5979f9b7a8b300da0145cea72518e598040f2e41bd"} Dec 06 08:23:45 crc kubenswrapper[4945]: I1206 08:23:45.263610 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1115-account-create-update-qkllx" event={"ID":"211109fe-4865-4082-ae88-8b4fb6facefa","Type":"ContainerStarted","Data":"fdfebbf4567440701848bb0f8067d0f66bf9b2305fb9307c25d6526e65fef417"} Dec 06 08:23:45 crc kubenswrapper[4945]: I1206 08:23:45.269596 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-fmsp8" event={"ID":"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98","Type":"ContainerStarted","Data":"7f8d820a4d63eb01023bf6e8bd8dbccc9109ff256c4b4e41191428d0db651904"} Dec 06 08:23:45 crc kubenswrapper[4945]: I1206 08:23:45.269846 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-fmsp8" event={"ID":"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98","Type":"ContainerStarted","Data":"3a9a028740ae5eda20842be8f1ec8e497719f1e26f9b2cbc51204ca7f888bb59"} Dec 06 08:23:45 crc kubenswrapper[4945]: I1206 08:23:45.317781 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-1115-account-create-update-qkllx" podStartSLOduration=1.31776415 podStartE2EDuration="1.31776415s" podCreationTimestamp="2025-12-06 08:23:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:23:45.296316476 +0000 UTC m=+5438.751177520" watchObservedRunningTime="2025-12-06 08:23:45.31776415 +0000 UTC m=+5438.772625194" Dec 06 08:23:45 crc kubenswrapper[4945]: I1206 08:23:45.320751 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-fmsp8" podStartSLOduration=2.32073849 podStartE2EDuration="2.32073849s" podCreationTimestamp="2025-12-06 08:23:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:23:45.312714055 +0000 UTC m=+5438.767575099" watchObservedRunningTime="2025-12-06 08:23:45.32073849 +0000 UTC m=+5438.775599534" Dec 06 08:23:46 crc kubenswrapper[4945]: I1206 08:23:46.279087 4945 generic.go:334] "Generic (PLEG): container finished" podID="211109fe-4865-4082-ae88-8b4fb6facefa" containerID="d1584e94d34ca3d0c00f1f5979f9b7a8b300da0145cea72518e598040f2e41bd" exitCode=0 Dec 06 08:23:46 crc kubenswrapper[4945]: I1206 08:23:46.279212 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1115-account-create-update-qkllx" event={"ID":"211109fe-4865-4082-ae88-8b4fb6facefa","Type":"ContainerDied","Data":"d1584e94d34ca3d0c00f1f5979f9b7a8b300da0145cea72518e598040f2e41bd"} Dec 06 08:23:46 crc kubenswrapper[4945]: I1206 08:23:46.281527 4945 generic.go:334] "Generic (PLEG): container finished" podID="7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98" containerID="7f8d820a4d63eb01023bf6e8bd8dbccc9109ff256c4b4e41191428d0db651904" exitCode=0 Dec 06 08:23:46 crc kubenswrapper[4945]: I1206 08:23:46.281614 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-fmsp8" event={"ID":"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98","Type":"ContainerDied","Data":"7f8d820a4d63eb01023bf6e8bd8dbccc9109ff256c4b4e41191428d0db651904"} Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.743892 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-fmsp8" Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.750635 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1115-account-create-update-qkllx" Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.871914 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/211109fe-4865-4082-ae88-8b4fb6facefa-operator-scripts\") pod \"211109fe-4865-4082-ae88-8b4fb6facefa\" (UID: \"211109fe-4865-4082-ae88-8b4fb6facefa\") " Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.871968 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-operator-scripts\") pod \"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98\" (UID: \"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98\") " Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.871993 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64n8q\" (UniqueName: \"kubernetes.io/projected/211109fe-4865-4082-ae88-8b4fb6facefa-kube-api-access-64n8q\") pod \"211109fe-4865-4082-ae88-8b4fb6facefa\" (UID: \"211109fe-4865-4082-ae88-8b4fb6facefa\") " Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.872080 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgsrw\" (UniqueName: \"kubernetes.io/projected/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-kube-api-access-kgsrw\") pod \"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98\" (UID: \"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98\") " Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.872906 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98" (UID: "7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.872950 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/211109fe-4865-4082-ae88-8b4fb6facefa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "211109fe-4865-4082-ae88-8b4fb6facefa" (UID: "211109fe-4865-4082-ae88-8b4fb6facefa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.878399 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/211109fe-4865-4082-ae88-8b4fb6facefa-kube-api-access-64n8q" (OuterVolumeSpecName: "kube-api-access-64n8q") pod "211109fe-4865-4082-ae88-8b4fb6facefa" (UID: "211109fe-4865-4082-ae88-8b4fb6facefa"). InnerVolumeSpecName "kube-api-access-64n8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.878498 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-kube-api-access-kgsrw" (OuterVolumeSpecName: "kube-api-access-kgsrw") pod "7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98" (UID: "7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98"). InnerVolumeSpecName "kube-api-access-kgsrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.974290 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgsrw\" (UniqueName: \"kubernetes.io/projected/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-kube-api-access-kgsrw\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.974324 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/211109fe-4865-4082-ae88-8b4fb6facefa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.974338 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:47 crc kubenswrapper[4945]: I1206 08:23:47.974347 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64n8q\" (UniqueName: \"kubernetes.io/projected/211109fe-4865-4082-ae88-8b4fb6facefa-kube-api-access-64n8q\") on node \"crc\" DevicePath \"\"" Dec 06 08:23:48 crc kubenswrapper[4945]: I1206 08:23:48.306547 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-fmsp8" event={"ID":"7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98","Type":"ContainerDied","Data":"3a9a028740ae5eda20842be8f1ec8e497719f1e26f9b2cbc51204ca7f888bb59"} Dec 06 08:23:48 crc kubenswrapper[4945]: I1206 08:23:48.306702 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-fmsp8" Dec 06 08:23:48 crc kubenswrapper[4945]: I1206 08:23:48.307131 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a9a028740ae5eda20842be8f1ec8e497719f1e26f9b2cbc51204ca7f888bb59" Dec 06 08:23:48 crc kubenswrapper[4945]: I1206 08:23:48.308352 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1115-account-create-update-qkllx" event={"ID":"211109fe-4865-4082-ae88-8b4fb6facefa","Type":"ContainerDied","Data":"fdfebbf4567440701848bb0f8067d0f66bf9b2305fb9307c25d6526e65fef417"} Dec 06 08:23:48 crc kubenswrapper[4945]: I1206 08:23:48.308404 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdfebbf4567440701848bb0f8067d0f66bf9b2305fb9307c25d6526e65fef417" Dec 06 08:23:48 crc kubenswrapper[4945]: I1206 08:23:48.308430 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1115-account-create-update-qkllx" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.341177 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-nzzsq"] Dec 06 08:23:49 crc kubenswrapper[4945]: E1206 08:23:49.341919 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="211109fe-4865-4082-ae88-8b4fb6facefa" containerName="mariadb-account-create-update" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.341934 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="211109fe-4865-4082-ae88-8b4fb6facefa" containerName="mariadb-account-create-update" Dec 06 08:23:49 crc kubenswrapper[4945]: E1206 08:23:49.341961 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98" containerName="mariadb-database-create" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.341969 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98" containerName="mariadb-database-create" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.342152 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98" containerName="mariadb-database-create" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.342182 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="211109fe-4865-4082-ae88-8b4fb6facefa" containerName="mariadb-account-create-update" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.342926 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.345411 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.345741 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xlhv2" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.347107 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.361361 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-nzzsq"] Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.413070 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-db-sync-config-data\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.413117 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/495f6a1b-e43d-436d-88ee-0f6fe915c51e-etc-machine-id\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.413138 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz56v\" (UniqueName: \"kubernetes.io/projected/495f6a1b-e43d-436d-88ee-0f6fe915c51e-kube-api-access-mz56v\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.413182 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-combined-ca-bundle\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.413219 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-config-data\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.413356 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-scripts\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.514415 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-scripts\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.514471 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-db-sync-config-data\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.514511 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/495f6a1b-e43d-436d-88ee-0f6fe915c51e-etc-machine-id\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.514531 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz56v\" (UniqueName: \"kubernetes.io/projected/495f6a1b-e43d-436d-88ee-0f6fe915c51e-kube-api-access-mz56v\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.514581 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-combined-ca-bundle\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.514615 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/495f6a1b-e43d-436d-88ee-0f6fe915c51e-etc-machine-id\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.514646 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-config-data\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.519995 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-db-sync-config-data\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.520055 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-combined-ca-bundle\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.522417 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-config-data\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.525216 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-scripts\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.531376 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz56v\" (UniqueName: \"kubernetes.io/projected/495f6a1b-e43d-436d-88ee-0f6fe915c51e-kube-api-access-mz56v\") pod \"cinder-db-sync-nzzsq\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:49 crc kubenswrapper[4945]: I1206 08:23:49.712810 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:23:50 crc kubenswrapper[4945]: I1206 08:23:50.162900 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-nzzsq"] Dec 06 08:23:50 crc kubenswrapper[4945]: W1206 08:23:50.163731 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod495f6a1b_e43d_436d_88ee_0f6fe915c51e.slice/crio-35478ed4a8384909aa72dfc140e0709fcf21225794ee42ea28a18c12e4fb2450 WatchSource:0}: Error finding container 35478ed4a8384909aa72dfc140e0709fcf21225794ee42ea28a18c12e4fb2450: Status 404 returned error can't find the container with id 35478ed4a8384909aa72dfc140e0709fcf21225794ee42ea28a18c12e4fb2450 Dec 06 08:23:50 crc kubenswrapper[4945]: I1206 08:23:50.334519 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nzzsq" event={"ID":"495f6a1b-e43d-436d-88ee-0f6fe915c51e","Type":"ContainerStarted","Data":"35478ed4a8384909aa72dfc140e0709fcf21225794ee42ea28a18c12e4fb2450"} Dec 06 08:23:52 crc kubenswrapper[4945]: I1206 08:23:52.953754 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:23:52 crc kubenswrapper[4945]: E1206 08:23:52.954612 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:24:07 crc kubenswrapper[4945]: I1206 08:24:07.953425 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:24:07 crc kubenswrapper[4945]: E1206 08:24:07.955541 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:24:10 crc kubenswrapper[4945]: E1206 08:24:10.576094 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:c3923531bcda0b0811b2d5053f189beb" Dec 06 08:24:10 crc kubenswrapper[4945]: E1206 08:24:10.576352 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:c3923531bcda0b0811b2d5053f189beb" Dec 06 08:24:10 crc kubenswrapper[4945]: E1206 08:24:10.576501 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mz56v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-nzzsq_openstack(495f6a1b-e43d-436d-88ee-0f6fe915c51e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 08:24:10 crc kubenswrapper[4945]: E1206 08:24:10.577843 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-nzzsq" podUID="495f6a1b-e43d-436d-88ee-0f6fe915c51e" Dec 06 08:24:11 crc kubenswrapper[4945]: E1206 08:24:11.540616 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:c3923531bcda0b0811b2d5053f189beb\\\"\"" pod="openstack/cinder-db-sync-nzzsq" podUID="495f6a1b-e43d-436d-88ee-0f6fe915c51e" Dec 06 08:24:18 crc kubenswrapper[4945]: I1206 08:24:18.953739 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:24:18 crc kubenswrapper[4945]: E1206 08:24:18.955679 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:24:24 crc kubenswrapper[4945]: I1206 08:24:24.656239 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nzzsq" event={"ID":"495f6a1b-e43d-436d-88ee-0f6fe915c51e","Type":"ContainerStarted","Data":"3645e18c9476c9157e4e2a7cbb91bd21bda41e44efddb7ac4d59eacc6f60cc11"} Dec 06 08:24:24 crc kubenswrapper[4945]: I1206 08:24:24.687759 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-nzzsq" podStartSLOduration=2.6908761 podStartE2EDuration="35.687737914s" podCreationTimestamp="2025-12-06 08:23:49 +0000 UTC" firstStartedPulling="2025-12-06 08:23:50.166028464 +0000 UTC m=+5443.620889518" lastFinishedPulling="2025-12-06 08:24:23.162890288 +0000 UTC m=+5476.617751332" observedRunningTime="2025-12-06 08:24:24.686676786 +0000 UTC m=+5478.141537920" watchObservedRunningTime="2025-12-06 08:24:24.687737914 +0000 UTC m=+5478.142598968" Dec 06 08:24:26 crc kubenswrapper[4945]: I1206 08:24:26.687343 4945 generic.go:334] "Generic (PLEG): container finished" podID="495f6a1b-e43d-436d-88ee-0f6fe915c51e" containerID="3645e18c9476c9157e4e2a7cbb91bd21bda41e44efddb7ac4d59eacc6f60cc11" exitCode=0 Dec 06 08:24:26 crc kubenswrapper[4945]: I1206 08:24:26.687353 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nzzsq" event={"ID":"495f6a1b-e43d-436d-88ee-0f6fe915c51e","Type":"ContainerDied","Data":"3645e18c9476c9157e4e2a7cbb91bd21bda41e44efddb7ac4d59eacc6f60cc11"} Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.110806 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.233070 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-scripts\") pod \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.233425 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/495f6a1b-e43d-436d-88ee-0f6fe915c51e-etc-machine-id\") pod \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.233452 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-config-data\") pod \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.233559 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-db-sync-config-data\") pod \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.233609 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-combined-ca-bundle\") pod \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.233612 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/495f6a1b-e43d-436d-88ee-0f6fe915c51e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "495f6a1b-e43d-436d-88ee-0f6fe915c51e" (UID: "495f6a1b-e43d-436d-88ee-0f6fe915c51e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.233721 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mz56v\" (UniqueName: \"kubernetes.io/projected/495f6a1b-e43d-436d-88ee-0f6fe915c51e-kube-api-access-mz56v\") pod \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\" (UID: \"495f6a1b-e43d-436d-88ee-0f6fe915c51e\") " Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.234074 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/495f6a1b-e43d-436d-88ee-0f6fe915c51e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.240031 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "495f6a1b-e43d-436d-88ee-0f6fe915c51e" (UID: "495f6a1b-e43d-436d-88ee-0f6fe915c51e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.240062 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-scripts" (OuterVolumeSpecName: "scripts") pod "495f6a1b-e43d-436d-88ee-0f6fe915c51e" (UID: "495f6a1b-e43d-436d-88ee-0f6fe915c51e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.240367 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/495f6a1b-e43d-436d-88ee-0f6fe915c51e-kube-api-access-mz56v" (OuterVolumeSpecName: "kube-api-access-mz56v") pod "495f6a1b-e43d-436d-88ee-0f6fe915c51e" (UID: "495f6a1b-e43d-436d-88ee-0f6fe915c51e"). InnerVolumeSpecName "kube-api-access-mz56v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.268924 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "495f6a1b-e43d-436d-88ee-0f6fe915c51e" (UID: "495f6a1b-e43d-436d-88ee-0f6fe915c51e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.299660 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-config-data" (OuterVolumeSpecName: "config-data") pod "495f6a1b-e43d-436d-88ee-0f6fe915c51e" (UID: "495f6a1b-e43d-436d-88ee-0f6fe915c51e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.336566 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mz56v\" (UniqueName: \"kubernetes.io/projected/495f6a1b-e43d-436d-88ee-0f6fe915c51e-kube-api-access-mz56v\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.336609 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.336623 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.336637 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.336652 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495f6a1b-e43d-436d-88ee-0f6fe915c51e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.710721 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nzzsq" event={"ID":"495f6a1b-e43d-436d-88ee-0f6fe915c51e","Type":"ContainerDied","Data":"35478ed4a8384909aa72dfc140e0709fcf21225794ee42ea28a18c12e4fb2450"} Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.710757 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35478ed4a8384909aa72dfc140e0709fcf21225794ee42ea28a18c12e4fb2450" Dec 06 08:24:28 crc kubenswrapper[4945]: I1206 08:24:28.710778 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nzzsq" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.033218 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64d4b844c-l5f9k"] Dec 06 08:24:29 crc kubenswrapper[4945]: E1206 08:24:29.033617 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495f6a1b-e43d-436d-88ee-0f6fe915c51e" containerName="cinder-db-sync" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.033628 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="495f6a1b-e43d-436d-88ee-0f6fe915c51e" containerName="cinder-db-sync" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.038742 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="495f6a1b-e43d-436d-88ee-0f6fe915c51e" containerName="cinder-db-sync" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.040103 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.047050 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64d4b844c-l5f9k"] Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.166221 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-dns-svc\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.166678 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-nb\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.166734 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlw87\" (UniqueName: \"kubernetes.io/projected/b1e39b0c-dfae-455c-9947-7af9b39a45f9-kube-api-access-jlw87\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.166751 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-config\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.166784 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-sb\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.175627 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.179205 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.182822 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.183364 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.183629 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.183803 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xlhv2" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.195509 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268456 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-dns-svc\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268520 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268545 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9f65\" (UniqueName: \"kubernetes.io/projected/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-kube-api-access-d9f65\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268576 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-nb\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268603 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-etc-machine-id\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268617 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-scripts\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268691 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268709 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlw87\" (UniqueName: \"kubernetes.io/projected/b1e39b0c-dfae-455c-9947-7af9b39a45f9-kube-api-access-jlw87\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268725 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-config\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268756 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data-custom\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268770 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-logs\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.268791 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-sb\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.269414 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-dns-svc\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.269685 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-sb\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.269831 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-nb\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.269908 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-config\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.305362 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlw87\" (UniqueName: \"kubernetes.io/projected/b1e39b0c-dfae-455c-9947-7af9b39a45f9-kube-api-access-jlw87\") pod \"dnsmasq-dns-64d4b844c-l5f9k\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.376746 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.376794 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9f65\" (UniqueName: \"kubernetes.io/projected/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-kube-api-access-d9f65\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.376853 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-etc-machine-id\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.376874 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-scripts\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.376922 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.376963 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data-custom\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.376985 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-logs\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.377274 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-etc-machine-id\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.378447 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-logs\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.381030 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data-custom\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.381991 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.383004 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.387793 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-scripts\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.397224 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9f65\" (UniqueName: \"kubernetes.io/projected/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-kube-api-access-d9f65\") pod \"cinder-api-0\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.405956 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.501248 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 08:24:29 crc kubenswrapper[4945]: I1206 08:24:29.954174 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:24:29 crc kubenswrapper[4945]: E1206 08:24:29.954953 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:24:30 crc kubenswrapper[4945]: I1206 08:24:30.020631 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64d4b844c-l5f9k"] Dec 06 08:24:30 crc kubenswrapper[4945]: W1206 08:24:30.028517 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1e39b0c_dfae_455c_9947_7af9b39a45f9.slice/crio-ebad84242f34f71d444cacd337c05c8be6560872a9b518550cbdc7d604438bd2 WatchSource:0}: Error finding container ebad84242f34f71d444cacd337c05c8be6560872a9b518550cbdc7d604438bd2: Status 404 returned error can't find the container with id ebad84242f34f71d444cacd337c05c8be6560872a9b518550cbdc7d604438bd2 Dec 06 08:24:30 crc kubenswrapper[4945]: I1206 08:24:30.155931 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 08:24:30 crc kubenswrapper[4945]: I1206 08:24:30.738139 4945 generic.go:334] "Generic (PLEG): container finished" podID="b1e39b0c-dfae-455c-9947-7af9b39a45f9" containerID="8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca" exitCode=0 Dec 06 08:24:30 crc kubenswrapper[4945]: I1206 08:24:30.738494 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" event={"ID":"b1e39b0c-dfae-455c-9947-7af9b39a45f9","Type":"ContainerDied","Data":"8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca"} Dec 06 08:24:30 crc kubenswrapper[4945]: I1206 08:24:30.738528 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" event={"ID":"b1e39b0c-dfae-455c-9947-7af9b39a45f9","Type":"ContainerStarted","Data":"ebad84242f34f71d444cacd337c05c8be6560872a9b518550cbdc7d604438bd2"} Dec 06 08:24:30 crc kubenswrapper[4945]: I1206 08:24:30.749392 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab","Type":"ContainerStarted","Data":"61f6c0dafb9b6cb843d164d96a7de89ba8013f1125c61d172c2a07ffa2df1387"} Dec 06 08:24:31 crc kubenswrapper[4945]: I1206 08:24:31.760178 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab","Type":"ContainerStarted","Data":"83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a"} Dec 06 08:24:31 crc kubenswrapper[4945]: I1206 08:24:31.760532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab","Type":"ContainerStarted","Data":"6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0"} Dec 06 08:24:31 crc kubenswrapper[4945]: I1206 08:24:31.760551 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 08:24:31 crc kubenswrapper[4945]: I1206 08:24:31.761996 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" event={"ID":"b1e39b0c-dfae-455c-9947-7af9b39a45f9","Type":"ContainerStarted","Data":"f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c"} Dec 06 08:24:31 crc kubenswrapper[4945]: I1206 08:24:31.762556 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:31 crc kubenswrapper[4945]: I1206 08:24:31.780713 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.780665874 podStartE2EDuration="2.780665874s" podCreationTimestamp="2025-12-06 08:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:24:31.77752169 +0000 UTC m=+5485.232382744" watchObservedRunningTime="2025-12-06 08:24:31.780665874 +0000 UTC m=+5485.235526918" Dec 06 08:24:39 crc kubenswrapper[4945]: I1206 08:24:39.407579 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:24:39 crc kubenswrapper[4945]: I1206 08:24:39.448002 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" podStartSLOduration=11.44796717 podStartE2EDuration="11.44796717s" podCreationTimestamp="2025-12-06 08:24:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:24:31.798055189 +0000 UTC m=+5485.252916233" watchObservedRunningTime="2025-12-06 08:24:39.44796717 +0000 UTC m=+5492.902828224" Dec 06 08:24:39 crc kubenswrapper[4945]: I1206 08:24:39.483985 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98f9d44d7-vnm9b"] Dec 06 08:24:39 crc kubenswrapper[4945]: I1206 08:24:39.484232 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" podUID="b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" containerName="dnsmasq-dns" containerID="cri-o://00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80" gracePeriod=10 Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.550403 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.703316 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-config\") pod \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.703454 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-dns-svc\") pod \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.703491 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbvc7\" (UniqueName: \"kubernetes.io/projected/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-kube-api-access-gbvc7\") pod \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.703544 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-nb\") pod \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.703572 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-sb\") pod \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\" (UID: \"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c\") " Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.710471 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-kube-api-access-gbvc7" (OuterVolumeSpecName: "kube-api-access-gbvc7") pod "b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" (UID: "b0a7f6b6-8b5c-42b6-b813-57fd600fa94c"). InnerVolumeSpecName "kube-api-access-gbvc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.752120 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-config" (OuterVolumeSpecName: "config") pod "b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" (UID: "b0a7f6b6-8b5c-42b6-b813-57fd600fa94c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.759812 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" (UID: "b0a7f6b6-8b5c-42b6-b813-57fd600fa94c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.761553 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" (UID: "b0a7f6b6-8b5c-42b6-b813-57fd600fa94c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.786740 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" (UID: "b0a7f6b6-8b5c-42b6-b813-57fd600fa94c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.805571 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.805631 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.805643 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbvc7\" (UniqueName: \"kubernetes.io/projected/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-kube-api-access-gbvc7\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.805654 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.805663 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.848802 4945 generic.go:334] "Generic (PLEG): container finished" podID="b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" containerID="00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80" exitCode=0 Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.848864 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" event={"ID":"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c","Type":"ContainerDied","Data":"00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80"} Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.848904 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" event={"ID":"b0a7f6b6-8b5c-42b6-b813-57fd600fa94c","Type":"ContainerDied","Data":"c5351124c621936b558b184a8e6a7be180bb3b02a7899f370f8dabbf0a070966"} Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.848932 4945 scope.go:117] "RemoveContainer" containerID="00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.849111 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98f9d44d7-vnm9b" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.873012 4945 scope.go:117] "RemoveContainer" containerID="531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.900390 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98f9d44d7-vnm9b"] Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.909799 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98f9d44d7-vnm9b"] Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.922855 4945 scope.go:117] "RemoveContainer" containerID="00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80" Dec 06 08:24:40 crc kubenswrapper[4945]: E1206 08:24:40.923431 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80\": container with ID starting with 00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80 not found: ID does not exist" containerID="00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.923461 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80"} err="failed to get container status \"00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80\": rpc error: code = NotFound desc = could not find container \"00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80\": container with ID starting with 00fc2ff11b1dde15ae1b5e23cb80a0991483144f8e2aaa2979caec01fa053a80 not found: ID does not exist" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.923484 4945 scope.go:117] "RemoveContainer" containerID="531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904" Dec 06 08:24:40 crc kubenswrapper[4945]: E1206 08:24:40.923670 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904\": container with ID starting with 531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904 not found: ID does not exist" containerID="531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.923690 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904"} err="failed to get container status \"531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904\": rpc error: code = NotFound desc = could not find container \"531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904\": container with ID starting with 531258b37051679f08ab4fc7f0c600989ca378ae6873ffa73b181bb78c40a904 not found: ID does not exist" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.936933 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q676j"] Dec 06 08:24:40 crc kubenswrapper[4945]: E1206 08:24:40.937512 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" containerName="dnsmasq-dns" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.937538 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" containerName="dnsmasq-dns" Dec 06 08:24:40 crc kubenswrapper[4945]: E1206 08:24:40.937571 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" containerName="init" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.937579 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" containerName="init" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.937840 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" containerName="dnsmasq-dns" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.939755 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.945959 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.946164 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="299c3108-fdd9-413b-a0cc-b751515e0ad0" containerName="nova-scheduler-scheduler" containerID="cri-o://a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" gracePeriod=30 Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.985207 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0a7f6b6-8b5c-42b6-b813-57fd600fa94c" path="/var/lib/kubelet/pods/b0a7f6b6-8b5c-42b6-b813-57fd600fa94c/volumes" Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.985887 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q676j"] Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.985909 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.985922 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.986095 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-log" containerID="cri-o://37fb7843aa59c006c625d448b556112a9a74ef369c01edcabd5789b8b7039f97" gracePeriod=30 Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.986255 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0412207c-21f7-4657-9213-ef4fba09c229" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://cedab13422fa411a54f0aa69ed9f44eab4068035dd24099951417aeba4b4c39b" gracePeriod=30 Dec 06 08:24:40 crc kubenswrapper[4945]: I1206 08:24:40.986474 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-metadata" containerID="cri-o://530de7d699d2067771b77ed69a84ff82d666a14f233167c207ba9271e9f2f8e5" gracePeriod=30 Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:40.994454 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:40.994675 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerName="nova-api-log" containerID="cri-o://5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff" gracePeriod=30 Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:40.994779 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerName="nova-api-api" containerID="cri-o://1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145" gracePeriod=30 Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.007130 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.007405 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="741d99e3-a6c8-4156-bb83-2ced76434e51" containerName="nova-cell0-conductor-conductor" containerID="cri-o://5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" gracePeriod=30 Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.009063 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-utilities\") pod \"redhat-operators-q676j\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.009237 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd6xj\" (UniqueName: \"kubernetes.io/projected/5fefb9ed-63da-4953-8b14-dd22ea4d339c-kube-api-access-kd6xj\") pod \"redhat-operators-q676j\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.009268 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-catalog-content\") pod \"redhat-operators-q676j\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.149449 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd6xj\" (UniqueName: \"kubernetes.io/projected/5fefb9ed-63da-4953-8b14-dd22ea4d339c-kube-api-access-kd6xj\") pod \"redhat-operators-q676j\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.149508 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-catalog-content\") pod \"redhat-operators-q676j\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.149551 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-utilities\") pod \"redhat-operators-q676j\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.154494 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-catalog-content\") pod \"redhat-operators-q676j\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.154708 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-utilities\") pod \"redhat-operators-q676j\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.234445 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd6xj\" (UniqueName: \"kubernetes.io/projected/5fefb9ed-63da-4953-8b14-dd22ea4d339c-kube-api-access-kd6xj\") pod \"redhat-operators-q676j\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:41 crc kubenswrapper[4945]: E1206 08:24:41.267471 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.268624 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:41 crc kubenswrapper[4945]: E1206 08:24:41.273399 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:24:41 crc kubenswrapper[4945]: E1206 08:24:41.280548 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:24:41 crc kubenswrapper[4945]: E1206 08:24:41.280630 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="299c3108-fdd9-413b-a0cc-b751515e0ad0" containerName="nova-scheduler-scheduler" Dec 06 08:24:41 crc kubenswrapper[4945]: E1206 08:24:41.659632 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:41 crc kubenswrapper[4945]: E1206 08:24:41.662555 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:41 crc kubenswrapper[4945]: E1206 08:24:41.667707 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:41 crc kubenswrapper[4945]: E1206 08:24:41.667748 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="741d99e3-a6c8-4156-bb83-2ced76434e51" containerName="nova-cell0-conductor-conductor" Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.844415 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q676j"] Dec 06 08:24:41 crc kubenswrapper[4945]: W1206 08:24:41.857704 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fefb9ed_63da_4953_8b14_dd22ea4d339c.slice/crio-59c4dd5262f75bdc964236a2bb7049846352cf94b5caad17abd7260a7b705d04 WatchSource:0}: Error finding container 59c4dd5262f75bdc964236a2bb7049846352cf94b5caad17abd7260a7b705d04: Status 404 returned error can't find the container with id 59c4dd5262f75bdc964236a2bb7049846352cf94b5caad17abd7260a7b705d04 Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.866681 4945 generic.go:334] "Generic (PLEG): container finished" podID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerID="5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff" exitCode=143 Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.866803 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8475348b-c57c-439a-ac05-1a71be4b29b8","Type":"ContainerDied","Data":"5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff"} Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.871852 4945 generic.go:334] "Generic (PLEG): container finished" podID="0412207c-21f7-4657-9213-ef4fba09c229" containerID="cedab13422fa411a54f0aa69ed9f44eab4068035dd24099951417aeba4b4c39b" exitCode=0 Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.871924 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0412207c-21f7-4657-9213-ef4fba09c229","Type":"ContainerDied","Data":"cedab13422fa411a54f0aa69ed9f44eab4068035dd24099951417aeba4b4c39b"} Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.882567 4945 generic.go:334] "Generic (PLEG): container finished" podID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerID="37fb7843aa59c006c625d448b556112a9a74ef369c01edcabd5789b8b7039f97" exitCode=143 Dec 06 08:24:41 crc kubenswrapper[4945]: I1206 08:24:41.882632 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6bb5ebc7-fa62-420c-b3a3-ff65df244f71","Type":"ContainerDied","Data":"37fb7843aa59c006c625d448b556112a9a74ef369c01edcabd5789b8b7039f97"} Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.209575 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.262859 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.404633 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-config-data\") pod \"0412207c-21f7-4657-9213-ef4fba09c229\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.404691 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-combined-ca-bundle\") pod \"0412207c-21f7-4657-9213-ef4fba09c229\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.404753 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bqdd\" (UniqueName: \"kubernetes.io/projected/0412207c-21f7-4657-9213-ef4fba09c229-kube-api-access-2bqdd\") pod \"0412207c-21f7-4657-9213-ef4fba09c229\" (UID: \"0412207c-21f7-4657-9213-ef4fba09c229\") " Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.414157 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0412207c-21f7-4657-9213-ef4fba09c229-kube-api-access-2bqdd" (OuterVolumeSpecName: "kube-api-access-2bqdd") pod "0412207c-21f7-4657-9213-ef4fba09c229" (UID: "0412207c-21f7-4657-9213-ef4fba09c229"). InnerVolumeSpecName "kube-api-access-2bqdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.442731 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-config-data" (OuterVolumeSpecName: "config-data") pod "0412207c-21f7-4657-9213-ef4fba09c229" (UID: "0412207c-21f7-4657-9213-ef4fba09c229"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.454434 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0412207c-21f7-4657-9213-ef4fba09c229" (UID: "0412207c-21f7-4657-9213-ef4fba09c229"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.506841 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.506879 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0412207c-21f7-4657-9213-ef4fba09c229-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.506889 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bqdd\" (UniqueName: \"kubernetes.io/projected/0412207c-21f7-4657-9213-ef4fba09c229-kube-api-access-2bqdd\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.912111 4945 generic.go:334] "Generic (PLEG): container finished" podID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerID="e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c" exitCode=0 Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.913594 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q676j" event={"ID":"5fefb9ed-63da-4953-8b14-dd22ea4d339c","Type":"ContainerDied","Data":"e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c"} Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.913658 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q676j" event={"ID":"5fefb9ed-63da-4953-8b14-dd22ea4d339c","Type":"ContainerStarted","Data":"59c4dd5262f75bdc964236a2bb7049846352cf94b5caad17abd7260a7b705d04"} Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.936215 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0412207c-21f7-4657-9213-ef4fba09c229","Type":"ContainerDied","Data":"d89055514ad5ebf818362cae9782a764e36cedeae7531a0223b5396c1a896fa1"} Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.936296 4945 scope.go:117] "RemoveContainer" containerID="cedab13422fa411a54f0aa69ed9f44eab4068035dd24099951417aeba4b4c39b" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.936473 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.957125 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:24:42 crc kubenswrapper[4945]: E1206 08:24:42.957400 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:24:42 crc kubenswrapper[4945]: I1206 08:24:42.998421 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.015205 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.028345 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 08:24:43 crc kubenswrapper[4945]: E1206 08:24:43.028790 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0412207c-21f7-4657-9213-ef4fba09c229" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.028807 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0412207c-21f7-4657-9213-ef4fba09c229" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.028997 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0412207c-21f7-4657-9213-ef4fba09c229" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.029680 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.034773 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.063923 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.116623 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtpjl\" (UniqueName: \"kubernetes.io/projected/f53d6629-9d10-4331-8f9f-ef26361c7e3b-kube-api-access-jtpjl\") pod \"nova-cell1-novncproxy-0\" (UID: \"f53d6629-9d10-4331-8f9f-ef26361c7e3b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.116723 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53d6629-9d10-4331-8f9f-ef26361c7e3b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f53d6629-9d10-4331-8f9f-ef26361c7e3b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.116875 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53d6629-9d10-4331-8f9f-ef26361c7e3b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f53d6629-9d10-4331-8f9f-ef26361c7e3b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.218544 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtpjl\" (UniqueName: \"kubernetes.io/projected/f53d6629-9d10-4331-8f9f-ef26361c7e3b-kube-api-access-jtpjl\") pod \"nova-cell1-novncproxy-0\" (UID: \"f53d6629-9d10-4331-8f9f-ef26361c7e3b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.218657 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53d6629-9d10-4331-8f9f-ef26361c7e3b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f53d6629-9d10-4331-8f9f-ef26361c7e3b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.218760 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53d6629-9d10-4331-8f9f-ef26361c7e3b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f53d6629-9d10-4331-8f9f-ef26361c7e3b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.232777 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53d6629-9d10-4331-8f9f-ef26361c7e3b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f53d6629-9d10-4331-8f9f-ef26361c7e3b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.235197 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53d6629-9d10-4331-8f9f-ef26361c7e3b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f53d6629-9d10-4331-8f9f-ef26361c7e3b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.237928 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtpjl\" (UniqueName: \"kubernetes.io/projected/f53d6629-9d10-4331-8f9f-ef26361c7e3b-kube-api-access-jtpjl\") pod \"nova-cell1-novncproxy-0\" (UID: \"f53d6629-9d10-4331-8f9f-ef26361c7e3b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.377590 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.838895 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.948020 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q676j" event={"ID":"5fefb9ed-63da-4953-8b14-dd22ea4d339c","Type":"ContainerStarted","Data":"7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb"} Dec 06 08:24:43 crc kubenswrapper[4945]: I1206 08:24:43.953139 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f53d6629-9d10-4331-8f9f-ef26361c7e3b","Type":"ContainerStarted","Data":"7208e139cca29c4ca5d8ef81947b75769b443577aba135d25e98fc8941ced842"} Dec 06 08:24:44 crc kubenswrapper[4945]: I1206 08:24:44.706169 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 08:24:44 crc kubenswrapper[4945]: I1206 08:24:44.706738 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="3896b278-e2a5-4576-a764-06cf62902109" containerName="nova-cell1-conductor-conductor" containerID="cri-o://4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf" gracePeriod=30 Dec 06 08:24:44 crc kubenswrapper[4945]: E1206 08:24:44.967361 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:44 crc kubenswrapper[4945]: E1206 08:24:44.982124 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:44 crc kubenswrapper[4945]: E1206 08:24:44.983873 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:44 crc kubenswrapper[4945]: E1206 08:24:44.983913 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="3896b278-e2a5-4576-a764-06cf62902109" containerName="nova-cell1-conductor-conductor" Dec 06 08:24:44 crc kubenswrapper[4945]: I1206 08:24:44.995178 4945 generic.go:334] "Generic (PLEG): container finished" podID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerID="7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb" exitCode=0 Dec 06 08:24:44 crc kubenswrapper[4945]: I1206 08:24:44.998810 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0412207c-21f7-4657-9213-ef4fba09c229" path="/var/lib/kubelet/pods/0412207c-21f7-4657-9213-ef4fba09c229/volumes" Dec 06 08:24:44 crc kubenswrapper[4945]: I1206 08:24:44.999630 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q676j" event={"ID":"5fefb9ed-63da-4953-8b14-dd22ea4d339c","Type":"ContainerDied","Data":"7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb"} Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.003791 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f53d6629-9d10-4331-8f9f-ef26361c7e3b","Type":"ContainerStarted","Data":"0f81113f6467055152d8d264de15f36b80418186f4c628651f040f9bbadbda07"} Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.044897 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.044878634 podStartE2EDuration="3.044878634s" podCreationTimestamp="2025-12-06 08:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:24:45.036408518 +0000 UTC m=+5498.491269582" watchObservedRunningTime="2025-12-06 08:24:45.044878634 +0000 UTC m=+5498.499739678" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.245797 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": dial tcp 10.217.1.69:8775: connect: connection refused" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.245790 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": dial tcp 10.217.1.69:8775: connect: connection refused" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.808359 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.881905 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cftd\" (UniqueName: \"kubernetes.io/projected/8475348b-c57c-439a-ac05-1a71be4b29b8-kube-api-access-6cftd\") pod \"8475348b-c57c-439a-ac05-1a71be4b29b8\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.882033 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-config-data\") pod \"8475348b-c57c-439a-ac05-1a71be4b29b8\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.882072 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-combined-ca-bundle\") pod \"8475348b-c57c-439a-ac05-1a71be4b29b8\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.882102 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8475348b-c57c-439a-ac05-1a71be4b29b8-logs\") pod \"8475348b-c57c-439a-ac05-1a71be4b29b8\" (UID: \"8475348b-c57c-439a-ac05-1a71be4b29b8\") " Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.883043 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8475348b-c57c-439a-ac05-1a71be4b29b8-logs" (OuterVolumeSpecName: "logs") pod "8475348b-c57c-439a-ac05-1a71be4b29b8" (UID: "8475348b-c57c-439a-ac05-1a71be4b29b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.904072 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8475348b-c57c-439a-ac05-1a71be4b29b8-kube-api-access-6cftd" (OuterVolumeSpecName: "kube-api-access-6cftd") pod "8475348b-c57c-439a-ac05-1a71be4b29b8" (UID: "8475348b-c57c-439a-ac05-1a71be4b29b8"). InnerVolumeSpecName "kube-api-access-6cftd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.921598 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8475348b-c57c-439a-ac05-1a71be4b29b8" (UID: "8475348b-c57c-439a-ac05-1a71be4b29b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.941428 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-config-data" (OuterVolumeSpecName: "config-data") pod "8475348b-c57c-439a-ac05-1a71be4b29b8" (UID: "8475348b-c57c-439a-ac05-1a71be4b29b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.985595 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cftd\" (UniqueName: \"kubernetes.io/projected/8475348b-c57c-439a-ac05-1a71be4b29b8-kube-api-access-6cftd\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.985630 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.985640 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8475348b-c57c-439a-ac05-1a71be4b29b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:45 crc kubenswrapper[4945]: I1206 08:24:45.985651 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8475348b-c57c-439a-ac05-1a71be4b29b8-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.031781 4945 generic.go:334] "Generic (PLEG): container finished" podID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerID="530de7d699d2067771b77ed69a84ff82d666a14f233167c207ba9271e9f2f8e5" exitCode=0 Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.031838 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6bb5ebc7-fa62-420c-b3a3-ff65df244f71","Type":"ContainerDied","Data":"530de7d699d2067771b77ed69a84ff82d666a14f233167c207ba9271e9f2f8e5"} Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.047594 4945 generic.go:334] "Generic (PLEG): container finished" podID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerID="1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145" exitCode=0 Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.047663 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8475348b-c57c-439a-ac05-1a71be4b29b8","Type":"ContainerDied","Data":"1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145"} Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.047691 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8475348b-c57c-439a-ac05-1a71be4b29b8","Type":"ContainerDied","Data":"1a56a562e47c14be1a7655a9274f333dc353bdc22be42a79f993a20f7f87dfe2"} Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.047708 4945 scope.go:117] "RemoveContainer" containerID="1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.047842 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.053601 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q676j" event={"ID":"5fefb9ed-63da-4953-8b14-dd22ea4d339c","Type":"ContainerStarted","Data":"c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7"} Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.076353 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q676j" podStartSLOduration=3.5335943309999998 podStartE2EDuration="6.076336821s" podCreationTimestamp="2025-12-06 08:24:40 +0000 UTC" firstStartedPulling="2025-12-06 08:24:42.91759601 +0000 UTC m=+5496.372457054" lastFinishedPulling="2025-12-06 08:24:45.4603385 +0000 UTC m=+5498.915199544" observedRunningTime="2025-12-06 08:24:46.073579247 +0000 UTC m=+5499.528440291" watchObservedRunningTime="2025-12-06 08:24:46.076336821 +0000 UTC m=+5499.531197865" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.241615 4945 scope.go:117] "RemoveContainer" containerID="5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.257626 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.259515 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.261235 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.264067 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.264194 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="299c3108-fdd9-413b-a0cc-b751515e0ad0" containerName="nova-scheduler-scheduler" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.273172 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.288726 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.289229 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerName="nova-api-api" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.289249 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerName="nova-api-api" Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.289295 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerName="nova-api-log" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.289302 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerName="nova-api-log" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.289518 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerName="nova-api-api" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.289540 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" containerName="nova-api-log" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.290604 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.293266 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.296754 4945 scope.go:117] "RemoveContainer" containerID="1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145" Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.297109 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145\": container with ID starting with 1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145 not found: ID does not exist" containerID="1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.297158 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145"} err="failed to get container status \"1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145\": rpc error: code = NotFound desc = could not find container \"1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145\": container with ID starting with 1c9a5d29aaedd48ec310604ca81a18baaa82e86892a4433b9ffd2a7457964145 not found: ID does not exist" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.297188 4945 scope.go:117] "RemoveContainer" containerID="5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff" Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.297917 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff\": container with ID starting with 5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff not found: ID does not exist" containerID="5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.297950 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff"} err="failed to get container status \"5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff\": rpc error: code = NotFound desc = could not find container \"5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff\": container with ID starting with 5f4963f8dfedf51bf2c3fb7d2a91dda7355ee32d51b99dd4d935a52b62d73dff not found: ID does not exist" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.316254 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.395674 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.395749 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlqdd\" (UniqueName: \"kubernetes.io/projected/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-kube-api-access-xlqdd\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.395802 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-logs\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.395911 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-config-data\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.459964 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.497182 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-config-data\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.497236 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.497311 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlqdd\" (UniqueName: \"kubernetes.io/projected/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-kube-api-access-xlqdd\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.497374 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-logs\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.497943 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-logs\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.502904 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.519215 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-config-data\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.524885 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlqdd\" (UniqueName: \"kubernetes.io/projected/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-kube-api-access-xlqdd\") pod \"nova-api-0\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.602676 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-logs\") pod \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.602748 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-config-data\") pod \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.602773 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgsfm\" (UniqueName: \"kubernetes.io/projected/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-kube-api-access-vgsfm\") pod \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.602802 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-combined-ca-bundle\") pod \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\" (UID: \"6bb5ebc7-fa62-420c-b3a3-ff65df244f71\") " Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.604033 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-logs" (OuterVolumeSpecName: "logs") pod "6bb5ebc7-fa62-420c-b3a3-ff65df244f71" (UID: "6bb5ebc7-fa62-420c-b3a3-ff65df244f71"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.606721 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-kube-api-access-vgsfm" (OuterVolumeSpecName: "kube-api-access-vgsfm") pod "6bb5ebc7-fa62-420c-b3a3-ff65df244f71" (UID: "6bb5ebc7-fa62-420c-b3a3-ff65df244f71"). InnerVolumeSpecName "kube-api-access-vgsfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.629855 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.636159 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-config-data" (OuterVolumeSpecName: "config-data") pod "6bb5ebc7-fa62-420c-b3a3-ff65df244f71" (UID: "6bb5ebc7-fa62-420c-b3a3-ff65df244f71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.648904 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bb5ebc7-fa62-420c-b3a3-ff65df244f71" (UID: "6bb5ebc7-fa62-420c-b3a3-ff65df244f71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.656617 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb is running failed: container process not found" containerID="5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.657120 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb is running failed: container process not found" containerID="5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.658539 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb is running failed: container process not found" containerID="5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:46 crc kubenswrapper[4945]: E1206 08:24:46.658575 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="741d99e3-a6c8-4156-bb83-2ced76434e51" containerName="nova-cell0-conductor-conductor" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.672101 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.704632 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.704667 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.704678 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgsfm\" (UniqueName: \"kubernetes.io/projected/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-kube-api-access-vgsfm\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.704690 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bb5ebc7-fa62-420c-b3a3-ff65df244f71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.805854 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-config-data\") pod \"741d99e3-a6c8-4156-bb83-2ced76434e51\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.806559 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-combined-ca-bundle\") pod \"741d99e3-a6c8-4156-bb83-2ced76434e51\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.807379 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcj7v\" (UniqueName: \"kubernetes.io/projected/741d99e3-a6c8-4156-bb83-2ced76434e51-kube-api-access-hcj7v\") pod \"741d99e3-a6c8-4156-bb83-2ced76434e51\" (UID: \"741d99e3-a6c8-4156-bb83-2ced76434e51\") " Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.811683 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/741d99e3-a6c8-4156-bb83-2ced76434e51-kube-api-access-hcj7v" (OuterVolumeSpecName: "kube-api-access-hcj7v") pod "741d99e3-a6c8-4156-bb83-2ced76434e51" (UID: "741d99e3-a6c8-4156-bb83-2ced76434e51"). InnerVolumeSpecName "kube-api-access-hcj7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.844447 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "741d99e3-a6c8-4156-bb83-2ced76434e51" (UID: "741d99e3-a6c8-4156-bb83-2ced76434e51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.863929 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-config-data" (OuterVolumeSpecName: "config-data") pod "741d99e3-a6c8-4156-bb83-2ced76434e51" (UID: "741d99e3-a6c8-4156-bb83-2ced76434e51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.909698 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.909737 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcj7v\" (UniqueName: \"kubernetes.io/projected/741d99e3-a6c8-4156-bb83-2ced76434e51-kube-api-access-hcj7v\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.909747 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/741d99e3-a6c8-4156-bb83-2ced76434e51-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:46 crc kubenswrapper[4945]: I1206 08:24:46.970195 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8475348b-c57c-439a-ac05-1a71be4b29b8" path="/var/lib/kubelet/pods/8475348b-c57c-439a-ac05-1a71be4b29b8/volumes" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.061940 4945 generic.go:334] "Generic (PLEG): container finished" podID="741d99e3-a6c8-4156-bb83-2ced76434e51" containerID="5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" exitCode=0 Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.062045 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.062048 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"741d99e3-a6c8-4156-bb83-2ced76434e51","Type":"ContainerDied","Data":"5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb"} Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.062230 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"741d99e3-a6c8-4156-bb83-2ced76434e51","Type":"ContainerDied","Data":"6444e684635d8d2694b5ac08930ed5a0516a0e6c23e16ab1e38889dc2e37db78"} Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.062267 4945 scope.go:117] "RemoveContainer" containerID="5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.065804 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.066198 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6bb5ebc7-fa62-420c-b3a3-ff65df244f71","Type":"ContainerDied","Data":"8cafdaae2eab9092135b9c4d9983fe2794d79f77866a7dd94b4ab60df6a3f7fb"} Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.089978 4945 scope.go:117] "RemoveContainer" containerID="5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" Dec 06 08:24:47 crc kubenswrapper[4945]: E1206 08:24:47.091805 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb\": container with ID starting with 5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb not found: ID does not exist" containerID="5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.091841 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb"} err="failed to get container status \"5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb\": rpc error: code = NotFound desc = could not find container \"5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb\": container with ID starting with 5a3b10f4d0628a053944c4c6c0eee4e46737748161c1865280787f1b6c7bc4eb not found: ID does not exist" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.091865 4945 scope.go:117] "RemoveContainer" containerID="530de7d699d2067771b77ed69a84ff82d666a14f233167c207ba9271e9f2f8e5" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.109351 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.135826 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.152592 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.188682 4945 scope.go:117] "RemoveContainer" containerID="37fb7843aa59c006c625d448b556112a9a74ef369c01edcabd5789b8b7039f97" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.190291 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.213239 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:24:47 crc kubenswrapper[4945]: E1206 08:24:47.214732 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="741d99e3-a6c8-4156-bb83-2ced76434e51" containerName="nova-cell0-conductor-conductor" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.214756 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="741d99e3-a6c8-4156-bb83-2ced76434e51" containerName="nova-cell0-conductor-conductor" Dec 06 08:24:47 crc kubenswrapper[4945]: E1206 08:24:47.214791 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-metadata" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.214801 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-metadata" Dec 06 08:24:47 crc kubenswrapper[4945]: E1206 08:24:47.214826 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-log" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.214835 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-log" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.216108 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-metadata" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.216141 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="741d99e3-a6c8-4156-bb83-2ced76434e51" containerName="nova-cell0-conductor-conductor" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.216176 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" containerName="nova-metadata-log" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.225225 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.231726 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.231901 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.233106 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.236467 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.250107 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.258042 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.269394 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.317308 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.317653 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz796\" (UniqueName: \"kubernetes.io/projected/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-kube-api-access-wz796\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.317687 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.317707 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.317731 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-config-data\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.317762 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-logs\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.318064 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvg7z\" (UniqueName: \"kubernetes.io/projected/a7804ea2-afec-4311-96f5-d5c95a13ed1f-kube-api-access-vvg7z\") pod \"nova-cell0-conductor-0\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.419694 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.419740 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.419769 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-config-data\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.419808 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-logs\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.419867 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvg7z\" (UniqueName: \"kubernetes.io/projected/a7804ea2-afec-4311-96f5-d5c95a13ed1f-kube-api-access-vvg7z\") pod \"nova-cell0-conductor-0\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.419927 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.419950 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz796\" (UniqueName: \"kubernetes.io/projected/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-kube-api-access-wz796\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.420682 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-logs\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.425972 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.433091 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.433550 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.436824 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-config-data\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.437701 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvg7z\" (UniqueName: \"kubernetes.io/projected/a7804ea2-afec-4311-96f5-d5c95a13ed1f-kube-api-access-vvg7z\") pod \"nova-cell0-conductor-0\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.439155 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz796\" (UniqueName: \"kubernetes.io/projected/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-kube-api-access-wz796\") pod \"nova-metadata-0\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.557902 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 08:24:47 crc kubenswrapper[4945]: I1206 08:24:47.675179 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:48 crc kubenswrapper[4945]: I1206 08:24:48.343705 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1907aea-7d79-4ed6-a35f-a8a501e4c03f","Type":"ContainerStarted","Data":"ddd55be65673d4137d1e955a25c2416a160f4212fbb22137802367313a580faa"} Dec 06 08:24:48 crc kubenswrapper[4945]: I1206 08:24:48.343754 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1907aea-7d79-4ed6-a35f-a8a501e4c03f","Type":"ContainerStarted","Data":"421e6fb693fb2a8656887c1a820052d1464dee0dd9b84fc91dc6ecf539a25ac9"} Dec 06 08:24:48 crc kubenswrapper[4945]: I1206 08:24:48.343765 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1907aea-7d79-4ed6-a35f-a8a501e4c03f","Type":"ContainerStarted","Data":"e2ff44ace196d1fc64d39241889b03c2175ca922301a7ae6f2e6ec74d57817dc"} Dec 06 08:24:48 crc kubenswrapper[4945]: I1206 08:24:48.378374 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 08:24:48 crc kubenswrapper[4945]: I1206 08:24:48.378752 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:48 crc kubenswrapper[4945]: I1206 08:24:48.381817 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.381794073 podStartE2EDuration="2.381794073s" podCreationTimestamp="2025-12-06 08:24:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:24:48.364931502 +0000 UTC m=+5501.819792556" watchObservedRunningTime="2025-12-06 08:24:48.381794073 +0000 UTC m=+5501.836655117" Dec 06 08:24:48 crc kubenswrapper[4945]: I1206 08:24:48.850792 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 08:24:48 crc kubenswrapper[4945]: I1206 08:24:48.969240 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bb5ebc7-fa62-420c-b3a3-ff65df244f71" path="/var/lib/kubelet/pods/6bb5ebc7-fa62-420c-b3a3-ff65df244f71/volumes" Dec 06 08:24:48 crc kubenswrapper[4945]: I1206 08:24:48.969848 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="741d99e3-a6c8-4156-bb83-2ced76434e51" path="/var/lib/kubelet/pods/741d99e3-a6c8-4156-bb83-2ced76434e51/volumes" Dec 06 08:24:49 crc kubenswrapper[4945]: I1206 08:24:49.354958 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7","Type":"ContainerStarted","Data":"d4960c3494a91d69c9f73b94d62151f7ed6516320a8f9a487843b140b1049866"} Dec 06 08:24:49 crc kubenswrapper[4945]: I1206 08:24:49.355288 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7","Type":"ContainerStarted","Data":"bedff4aff87aa6756de152dac23aeef4b9a193a18406e52672a49e628dd6fdd3"} Dec 06 08:24:49 crc kubenswrapper[4945]: I1206 08:24:49.355303 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7","Type":"ContainerStarted","Data":"e4afa7894da57726fcb7929f4d927005caca2809ac36e650867630fc3e9c0591"} Dec 06 08:24:49 crc kubenswrapper[4945]: I1206 08:24:49.357673 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a7804ea2-afec-4311-96f5-d5c95a13ed1f","Type":"ContainerStarted","Data":"9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3"} Dec 06 08:24:49 crc kubenswrapper[4945]: I1206 08:24:49.357701 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a7804ea2-afec-4311-96f5-d5c95a13ed1f","Type":"ContainerStarted","Data":"447a4fc05de903eaec22613235f611cc48e2a5c92a232cd5cfd6c75e9e620882"} Dec 06 08:24:49 crc kubenswrapper[4945]: I1206 08:24:49.357847 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:49 crc kubenswrapper[4945]: I1206 08:24:49.379873 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.379854776 podStartE2EDuration="2.379854776s" podCreationTimestamp="2025-12-06 08:24:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:24:49.371975145 +0000 UTC m=+5502.826836189" watchObservedRunningTime="2025-12-06 08:24:49.379854776 +0000 UTC m=+5502.834715820" Dec 06 08:24:49 crc kubenswrapper[4945]: I1206 08:24:49.399933 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.399914103 podStartE2EDuration="2.399914103s" podCreationTimestamp="2025-12-06 08:24:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:24:49.392023741 +0000 UTC m=+5502.846884795" watchObservedRunningTime="2025-12-06 08:24:49.399914103 +0000 UTC m=+5502.854775147" Dec 06 08:24:49 crc kubenswrapper[4945]: E1206 08:24:49.960672 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf is running failed: container process not found" containerID="4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:49 crc kubenswrapper[4945]: E1206 08:24:49.961015 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf is running failed: container process not found" containerID="4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:49 crc kubenswrapper[4945]: E1206 08:24:49.961346 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf is running failed: container process not found" containerID="4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 08:24:49 crc kubenswrapper[4945]: E1206 08:24:49.961384 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="3896b278-e2a5-4576-a764-06cf62902109" containerName="nova-cell1-conductor-conductor" Dec 06 08:24:51 crc kubenswrapper[4945]: E1206 08:24:51.229506 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d is running failed: container process not found" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:24:51 crc kubenswrapper[4945]: E1206 08:24:51.230384 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d is running failed: container process not found" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:24:51 crc kubenswrapper[4945]: E1206 08:24:51.233016 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d is running failed: container process not found" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 08:24:51 crc kubenswrapper[4945]: E1206 08:24:51.233086 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="299c3108-fdd9-413b-a0cc-b751515e0ad0" containerName="nova-scheduler-scheduler" Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.268980 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.269570 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.377512 4945 generic.go:334] "Generic (PLEG): container finished" podID="3896b278-e2a5-4576-a764-06cf62902109" containerID="4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf" exitCode=0 Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.377594 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"3896b278-e2a5-4576-a764-06cf62902109","Type":"ContainerDied","Data":"4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf"} Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.383958 4945 generic.go:334] "Generic (PLEG): container finished" podID="299c3108-fdd9-413b-a0cc-b751515e0ad0" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" exitCode=0 Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.384812 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"299c3108-fdd9-413b-a0cc-b751515e0ad0","Type":"ContainerDied","Data":"a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d"} Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.714061 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.886705 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbpl2\" (UniqueName: \"kubernetes.io/projected/3896b278-e2a5-4576-a764-06cf62902109-kube-api-access-xbpl2\") pod \"3896b278-e2a5-4576-a764-06cf62902109\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.886933 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-combined-ca-bundle\") pod \"3896b278-e2a5-4576-a764-06cf62902109\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.887041 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-config-data\") pod \"3896b278-e2a5-4576-a764-06cf62902109\" (UID: \"3896b278-e2a5-4576-a764-06cf62902109\") " Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.904067 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3896b278-e2a5-4576-a764-06cf62902109-kube-api-access-xbpl2" (OuterVolumeSpecName: "kube-api-access-xbpl2") pod "3896b278-e2a5-4576-a764-06cf62902109" (UID: "3896b278-e2a5-4576-a764-06cf62902109"). InnerVolumeSpecName "kube-api-access-xbpl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.922886 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-config-data" (OuterVolumeSpecName: "config-data") pod "3896b278-e2a5-4576-a764-06cf62902109" (UID: "3896b278-e2a5-4576-a764-06cf62902109"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.935541 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3896b278-e2a5-4576-a764-06cf62902109" (UID: "3896b278-e2a5-4576-a764-06cf62902109"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.991180 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.991213 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbpl2\" (UniqueName: \"kubernetes.io/projected/3896b278-e2a5-4576-a764-06cf62902109-kube-api-access-xbpl2\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:51 crc kubenswrapper[4945]: I1206 08:24:51.991223 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3896b278-e2a5-4576-a764-06cf62902109-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.057772 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.193499 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-combined-ca-bundle\") pod \"299c3108-fdd9-413b-a0cc-b751515e0ad0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.193646 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-config-data\") pod \"299c3108-fdd9-413b-a0cc-b751515e0ad0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.193788 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpj2h\" (UniqueName: \"kubernetes.io/projected/299c3108-fdd9-413b-a0cc-b751515e0ad0-kube-api-access-kpj2h\") pod \"299c3108-fdd9-413b-a0cc-b751515e0ad0\" (UID: \"299c3108-fdd9-413b-a0cc-b751515e0ad0\") " Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.200451 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/299c3108-fdd9-413b-a0cc-b751515e0ad0-kube-api-access-kpj2h" (OuterVolumeSpecName: "kube-api-access-kpj2h") pod "299c3108-fdd9-413b-a0cc-b751515e0ad0" (UID: "299c3108-fdd9-413b-a0cc-b751515e0ad0"). InnerVolumeSpecName "kube-api-access-kpj2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.224462 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-config-data" (OuterVolumeSpecName: "config-data") pod "299c3108-fdd9-413b-a0cc-b751515e0ad0" (UID: "299c3108-fdd9-413b-a0cc-b751515e0ad0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.227550 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "299c3108-fdd9-413b-a0cc-b751515e0ad0" (UID: "299c3108-fdd9-413b-a0cc-b751515e0ad0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.297652 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpj2h\" (UniqueName: \"kubernetes.io/projected/299c3108-fdd9-413b-a0cc-b751515e0ad0-kube-api-access-kpj2h\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.297689 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.297704 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/299c3108-fdd9-413b-a0cc-b751515e0ad0-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.321510 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q676j" podUID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerName="registry-server" probeResult="failure" output=< Dec 06 08:24:52 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 08:24:52 crc kubenswrapper[4945]: > Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.393598 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"299c3108-fdd9-413b-a0cc-b751515e0ad0","Type":"ContainerDied","Data":"e0cad7e62909687eee6c985804ff5f031778fd1e23f1a55848f48880b96c9432"} Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.393680 4945 scope.go:117] "RemoveContainer" containerID="a7692eb8eb07b9c9ec30fe47573aa61ea5dce629a19232785b6a487435d4266d" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.393618 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.395342 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"3896b278-e2a5-4576-a764-06cf62902109","Type":"ContainerDied","Data":"0c963f5b0a7ab016bd769f8a1b43eb3e84f2092dc305e2df4aae08c172bc0a4c"} Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.395409 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.416436 4945 scope.go:117] "RemoveContainer" containerID="4c7cfbaf52862f8971ceb67a8556b93e80bfd0e6aa716b9d71c8ee423a44dabf" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.452938 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.464569 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.476120 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.488624 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.499265 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 08:24:52 crc kubenswrapper[4945]: E1206 08:24:52.499807 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3896b278-e2a5-4576-a764-06cf62902109" containerName="nova-cell1-conductor-conductor" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.499827 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3896b278-e2a5-4576-a764-06cf62902109" containerName="nova-cell1-conductor-conductor" Dec 06 08:24:52 crc kubenswrapper[4945]: E1206 08:24:52.499845 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="299c3108-fdd9-413b-a0cc-b751515e0ad0" containerName="nova-scheduler-scheduler" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.499855 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="299c3108-fdd9-413b-a0cc-b751515e0ad0" containerName="nova-scheduler-scheduler" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.500098 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3896b278-e2a5-4576-a764-06cf62902109" containerName="nova-cell1-conductor-conductor" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.500125 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="299c3108-fdd9-413b-a0cc-b751515e0ad0" containerName="nova-scheduler-scheduler" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.500896 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.503559 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.509460 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.522657 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.524182 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.526646 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.534394 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.562340 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.562387 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.703860 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-config-data\") pod \"nova-scheduler-0\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.703917 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.704035 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg9bl\" (UniqueName: \"kubernetes.io/projected/19af6ee2-c8a2-446b-a151-fe19310fa749-kube-api-access-rg9bl\") pod \"nova-scheduler-0\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.704061 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.704149 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tblxm\" (UniqueName: \"kubernetes.io/projected/9e769c4b-cc65-4469-a585-5d7423bb3f87-kube-api-access-tblxm\") pod \"nova-cell1-conductor-0\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.704251 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.805820 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.805976 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg9bl\" (UniqueName: \"kubernetes.io/projected/19af6ee2-c8a2-446b-a151-fe19310fa749-kube-api-access-rg9bl\") pod \"nova-scheduler-0\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.806005 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.806034 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tblxm\" (UniqueName: \"kubernetes.io/projected/9e769c4b-cc65-4469-a585-5d7423bb3f87-kube-api-access-tblxm\") pod \"nova-cell1-conductor-0\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.806093 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.806146 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-config-data\") pod \"nova-scheduler-0\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.810604 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.811210 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.811337 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.818001 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-config-data\") pod \"nova-scheduler-0\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.822972 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tblxm\" (UniqueName: \"kubernetes.io/projected/9e769c4b-cc65-4469-a585-5d7423bb3f87-kube-api-access-tblxm\") pod \"nova-cell1-conductor-0\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.823508 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg9bl\" (UniqueName: \"kubernetes.io/projected/19af6ee2-c8a2-446b-a151-fe19310fa749-kube-api-access-rg9bl\") pod \"nova-scheduler-0\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.839865 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.969873 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="299c3108-fdd9-413b-a0cc-b751515e0ad0" path="/var/lib/kubelet/pods/299c3108-fdd9-413b-a0cc-b751515e0ad0/volumes" Dec 06 08:24:52 crc kubenswrapper[4945]: I1206 08:24:52.970974 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3896b278-e2a5-4576-a764-06cf62902109" path="/var/lib/kubelet/pods/3896b278-e2a5-4576-a764-06cf62902109/volumes" Dec 06 08:24:53 crc kubenswrapper[4945]: I1206 08:24:53.118553 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:53 crc kubenswrapper[4945]: I1206 08:24:53.317113 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 08:24:53 crc kubenswrapper[4945]: W1206 08:24:53.323722 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19af6ee2_c8a2_446b_a151_fe19310fa749.slice/crio-6f1f0af4f152ac1e76650df0677c614e8cfb732fcb1154f0da997737f892390b WatchSource:0}: Error finding container 6f1f0af4f152ac1e76650df0677c614e8cfb732fcb1154f0da997737f892390b: Status 404 returned error can't find the container with id 6f1f0af4f152ac1e76650df0677c614e8cfb732fcb1154f0da997737f892390b Dec 06 08:24:53 crc kubenswrapper[4945]: I1206 08:24:53.379706 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:53 crc kubenswrapper[4945]: I1206 08:24:53.397954 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:53 crc kubenswrapper[4945]: I1206 08:24:53.407605 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19af6ee2-c8a2-446b-a151-fe19310fa749","Type":"ContainerStarted","Data":"6f1f0af4f152ac1e76650df0677c614e8cfb732fcb1154f0da997737f892390b"} Dec 06 08:24:53 crc kubenswrapper[4945]: I1206 08:24:53.424373 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 06 08:24:53 crc kubenswrapper[4945]: I1206 08:24:53.640071 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 08:24:54 crc kubenswrapper[4945]: I1206 08:24:54.423947 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19af6ee2-c8a2-446b-a151-fe19310fa749","Type":"ContainerStarted","Data":"cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1"} Dec 06 08:24:54 crc kubenswrapper[4945]: I1206 08:24:54.426015 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9e769c4b-cc65-4469-a585-5d7423bb3f87","Type":"ContainerStarted","Data":"187e17d0582fac38c827610ffc14e5cef0886eb9335d4686ad45f6f3b651c1b2"} Dec 06 08:24:54 crc kubenswrapper[4945]: I1206 08:24:54.426069 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9e769c4b-cc65-4469-a585-5d7423bb3f87","Type":"ContainerStarted","Data":"a1fc1d704317c11bcb666d1c69af66afd7bd8e437868404c30a5f2144839de0d"} Dec 06 08:24:54 crc kubenswrapper[4945]: I1206 08:24:54.426276 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 06 08:24:54 crc kubenswrapper[4945]: I1206 08:24:54.447946 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.447923001 podStartE2EDuration="2.447923001s" podCreationTimestamp="2025-12-06 08:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:24:54.439680281 +0000 UTC m=+5507.894541325" watchObservedRunningTime="2025-12-06 08:24:54.447923001 +0000 UTC m=+5507.902784055" Dec 06 08:24:54 crc kubenswrapper[4945]: I1206 08:24:54.460791 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.4607753949999998 podStartE2EDuration="2.460775395s" podCreationTimestamp="2025-12-06 08:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:24:54.45198942 +0000 UTC m=+5507.906850464" watchObservedRunningTime="2025-12-06 08:24:54.460775395 +0000 UTC m=+5507.915636429" Dec 06 08:24:56 crc kubenswrapper[4945]: I1206 08:24:56.630903 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 08:24:56 crc kubenswrapper[4945]: I1206 08:24:56.631150 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 08:24:57 crc kubenswrapper[4945]: I1206 08:24:57.559123 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 08:24:57 crc kubenswrapper[4945]: I1206 08:24:57.559174 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 08:24:57 crc kubenswrapper[4945]: I1206 08:24:57.712451 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.80:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:24:57 crc kubenswrapper[4945]: I1206 08:24:57.712523 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.80:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:24:57 crc kubenswrapper[4945]: I1206 08:24:57.721579 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 06 08:24:57 crc kubenswrapper[4945]: I1206 08:24:57.840915 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 08:24:57 crc kubenswrapper[4945]: I1206 08:24:57.954332 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:24:57 crc kubenswrapper[4945]: E1206 08:24:57.954741 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:24:58 crc kubenswrapper[4945]: I1206 08:24:58.599643 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:24:58 crc kubenswrapper[4945]: I1206 08:24:58.599708 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 08:25:01 crc kubenswrapper[4945]: I1206 08:25:01.310374 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:25:01 crc kubenswrapper[4945]: I1206 08:25:01.367093 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:25:01 crc kubenswrapper[4945]: I1206 08:25:01.548492 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q676j"] Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.433171 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.435123 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.437363 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.445336 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.504749 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q676j" podUID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerName="registry-server" containerID="cri-o://c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7" gracePeriod=2 Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.580935 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.581072 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.581200 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.581263 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bm2p\" (UniqueName: \"kubernetes.io/projected/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-kube-api-access-5bm2p\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.581403 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-scripts\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.581474 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.682768 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-scripts\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.682834 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.682901 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.682944 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.683021 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.683076 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bm2p\" (UniqueName: \"kubernetes.io/projected/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-kube-api-access-5bm2p\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.683960 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.691349 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.693985 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.694268 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-scripts\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.694638 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.706957 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bm2p\" (UniqueName: \"kubernetes.io/projected/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-kube-api-access-5bm2p\") pod \"cinder-scheduler-0\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.752407 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.842115 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 08:25:02 crc kubenswrapper[4945]: I1206 08:25:02.915429 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.013648 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.155116 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.191915 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-catalog-content\") pod \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.192038 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kd6xj\" (UniqueName: \"kubernetes.io/projected/5fefb9ed-63da-4953-8b14-dd22ea4d339c-kube-api-access-kd6xj\") pod \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.192090 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-utilities\") pod \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\" (UID: \"5fefb9ed-63da-4953-8b14-dd22ea4d339c\") " Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.193245 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-utilities" (OuterVolumeSpecName: "utilities") pod "5fefb9ed-63da-4953-8b14-dd22ea4d339c" (UID: "5fefb9ed-63da-4953-8b14-dd22ea4d339c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.197329 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fefb9ed-63da-4953-8b14-dd22ea4d339c-kube-api-access-kd6xj" (OuterVolumeSpecName: "kube-api-access-kd6xj") pod "5fefb9ed-63da-4953-8b14-dd22ea4d339c" (UID: "5fefb9ed-63da-4953-8b14-dd22ea4d339c"). InnerVolumeSpecName "kube-api-access-kd6xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.276756 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 08:25:03 crc kubenswrapper[4945]: W1206 08:25:03.277199 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod788bc4d9_95d2_4b50_bf1b_3d5c9f844f60.slice/crio-7e98b0dc3d6e9251165b34ff57cec947d6cc04651718638999486fb97017d2cd WatchSource:0}: Error finding container 7e98b0dc3d6e9251165b34ff57cec947d6cc04651718638999486fb97017d2cd: Status 404 returned error can't find the container with id 7e98b0dc3d6e9251165b34ff57cec947d6cc04651718638999486fb97017d2cd Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.293950 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kd6xj\" (UniqueName: \"kubernetes.io/projected/5fefb9ed-63da-4953-8b14-dd22ea4d339c-kube-api-access-kd6xj\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.293978 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.295696 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fefb9ed-63da-4953-8b14-dd22ea4d339c" (UID: "5fefb9ed-63da-4953-8b14-dd22ea4d339c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.395846 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fefb9ed-63da-4953-8b14-dd22ea4d339c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.523005 4945 generic.go:334] "Generic (PLEG): container finished" podID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerID="c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7" exitCode=0 Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.523444 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q676j" event={"ID":"5fefb9ed-63da-4953-8b14-dd22ea4d339c","Type":"ContainerDied","Data":"c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7"} Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.524213 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q676j" event={"ID":"5fefb9ed-63da-4953-8b14-dd22ea4d339c","Type":"ContainerDied","Data":"59c4dd5262f75bdc964236a2bb7049846352cf94b5caad17abd7260a7b705d04"} Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.523469 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q676j" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.524304 4945 scope.go:117] "RemoveContainer" containerID="c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.529549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60","Type":"ContainerStarted","Data":"7e98b0dc3d6e9251165b34ff57cec947d6cc04651718638999486fb97017d2cd"} Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.549827 4945 scope.go:117] "RemoveContainer" containerID="7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.568845 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q676j"] Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.576421 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q676j"] Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.577028 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.617313 4945 scope.go:117] "RemoveContainer" containerID="e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.800941 4945 scope.go:117] "RemoveContainer" containerID="c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7" Dec 06 08:25:03 crc kubenswrapper[4945]: E1206 08:25:03.801913 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7\": container with ID starting with c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7 not found: ID does not exist" containerID="c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.802032 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7"} err="failed to get container status \"c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7\": rpc error: code = NotFound desc = could not find container \"c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7\": container with ID starting with c23ca62d530db71a5ce2b797390c64a393893c41209ccee46c27d15882b891f7 not found: ID does not exist" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.802067 4945 scope.go:117] "RemoveContainer" containerID="7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb" Dec 06 08:25:03 crc kubenswrapper[4945]: E1206 08:25:03.802597 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb\": container with ID starting with 7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb not found: ID does not exist" containerID="7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.802714 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb"} err="failed to get container status \"7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb\": rpc error: code = NotFound desc = could not find container \"7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb\": container with ID starting with 7c3052e820490c84b01a838cb4c9f241a8033ea14c242ac4d4a8f2b9b16b17cb not found: ID does not exist" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.802824 4945 scope.go:117] "RemoveContainer" containerID="e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c" Dec 06 08:25:03 crc kubenswrapper[4945]: E1206 08:25:03.803240 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c\": container with ID starting with e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c not found: ID does not exist" containerID="e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c" Dec 06 08:25:03 crc kubenswrapper[4945]: I1206 08:25:03.803270 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c"} err="failed to get container status \"e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c\": rpc error: code = NotFound desc = could not find container \"e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c\": container with ID starting with e8947ed95889b52dae10e1ac8ee2004ff296833e23fd08508288cb3c46241c2c not found: ID does not exist" Dec 06 08:25:04 crc kubenswrapper[4945]: I1206 08:25:04.565399 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60","Type":"ContainerStarted","Data":"10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88"} Dec 06 08:25:04 crc kubenswrapper[4945]: I1206 08:25:04.618529 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 08:25:04 crc kubenswrapper[4945]: I1206 08:25:04.619125 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" containerName="cinder-api-log" containerID="cri-o://6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0" gracePeriod=30 Dec 06 08:25:04 crc kubenswrapper[4945]: I1206 08:25:04.619208 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" containerName="cinder-api" containerID="cri-o://83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a" gracePeriod=30 Dec 06 08:25:04 crc kubenswrapper[4945]: I1206 08:25:04.964450 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" path="/var/lib/kubelet/pods/5fefb9ed-63da-4953-8b14-dd22ea4d339c/volumes" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.124138 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 06 08:25:05 crc kubenswrapper[4945]: E1206 08:25:05.124603 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerName="extract-utilities" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.124623 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerName="extract-utilities" Dec 06 08:25:05 crc kubenswrapper[4945]: E1206 08:25:05.124647 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerName="extract-content" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.124655 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerName="extract-content" Dec 06 08:25:05 crc kubenswrapper[4945]: E1206 08:25:05.124692 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerName="registry-server" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.124701 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerName="registry-server" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.124940 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fefb9ed-63da-4953-8b14-dd22ea4d339c" containerName="registry-server" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.126315 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.132128 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.150073 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.226578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227066 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227149 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227263 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227366 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227444 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnc8f\" (UniqueName: \"kubernetes.io/projected/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-kube-api-access-jnc8f\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227626 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227719 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227801 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227871 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.227947 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.228028 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-run\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.228106 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.228216 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.228332 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.329400 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.329724 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.329830 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.329919 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.329846 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.329985 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330151 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330146 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330309 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330396 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330319 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330182 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330364 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330518 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnc8f\" (UniqueName: \"kubernetes.io/projected/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-kube-api-access-jnc8f\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330710 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330818 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330908 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.330971 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.331071 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.331146 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.331187 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.331118 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.331383 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.331505 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-run\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.331562 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-run\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.335475 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.335525 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.335885 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.344166 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.346795 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.347788 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnc8f\" (UniqueName: \"kubernetes.io/projected/a15a0964-44c0-4a17-9f3e-42a5ed41f4e9-kube-api-access-jnc8f\") pod \"cinder-volume-volume1-0\" (UID: \"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9\") " pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.459756 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.581022 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60","Type":"ContainerStarted","Data":"7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db"} Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.588376 4945 generic.go:334] "Generic (PLEG): container finished" podID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" containerID="6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0" exitCode=143 Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.588426 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab","Type":"ContainerDied","Data":"6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0"} Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.608253 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.268613346 podStartE2EDuration="3.608234662s" podCreationTimestamp="2025-12-06 08:25:02 +0000 UTC" firstStartedPulling="2025-12-06 08:25:03.27960943 +0000 UTC m=+5516.734470474" lastFinishedPulling="2025-12-06 08:25:03.619230736 +0000 UTC m=+5517.074091790" observedRunningTime="2025-12-06 08:25:05.605131209 +0000 UTC m=+5519.059992253" watchObservedRunningTime="2025-12-06 08:25:05.608234662 +0000 UTC m=+5519.063095706" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.942040 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.944701 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.946344 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 06 08:25:05 crc kubenswrapper[4945]: I1206 08:25:05.962057 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.045223 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.045675 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.045749 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/737bd8cb-ebc6-441f-9f6c-0f13e356232c-ceph\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.045846 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.045918 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-lib-modules\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.045993 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-run\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.046087 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xblfj\" (UniqueName: \"kubernetes.io/projected/737bd8cb-ebc6-441f-9f6c-0f13e356232c-kube-api-access-xblfj\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.046151 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.046368 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-config-data\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.046444 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.046528 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.046603 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-sys\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.046689 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.046766 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.046828 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-scripts\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.046924 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-dev\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.081464 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 06 08:25:06 crc kubenswrapper[4945]: W1206 08:25:06.083229 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda15a0964_44c0_4a17_9f3e_42a5ed41f4e9.slice/crio-f5e4e69b21b7721e7076d3062f2f2ebaf3f2bac8ac67be53962baf4eb97108ae WatchSource:0}: Error finding container f5e4e69b21b7721e7076d3062f2f2ebaf3f2bac8ac67be53962baf4eb97108ae: Status 404 returned error can't find the container with id f5e4e69b21b7721e7076d3062f2f2ebaf3f2bac8ac67be53962baf4eb97108ae Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.148643 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xblfj\" (UniqueName: \"kubernetes.io/projected/737bd8cb-ebc6-441f-9f6c-0f13e356232c-kube-api-access-xblfj\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.148693 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.148776 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-config-data\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.148810 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.148846 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.148872 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-sys\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.148904 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.148938 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.148961 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-scripts\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.148988 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-dev\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149020 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149045 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149066 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/737bd8cb-ebc6-441f-9f6c-0f13e356232c-ceph\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149103 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149128 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-lib-modules\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149159 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-run\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149268 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-run\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149496 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-dev\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149525 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149506 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-lib-modules\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149559 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149596 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-sys\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149623 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149661 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149684 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.149790 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/737bd8cb-ebc6-441f-9f6c-0f13e356232c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.155184 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-scripts\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.155595 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/737bd8cb-ebc6-441f-9f6c-0f13e356232c-ceph\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.156103 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.156987 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.181156 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xblfj\" (UniqueName: \"kubernetes.io/projected/737bd8cb-ebc6-441f-9f6c-0f13e356232c-kube-api-access-xblfj\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.182142 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/737bd8cb-ebc6-441f-9f6c-0f13e356232c-config-data\") pod \"cinder-backup-0\" (UID: \"737bd8cb-ebc6-441f-9f6c-0f13e356232c\") " pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.299109 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.632969 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9","Type":"ContainerStarted","Data":"f5e4e69b21b7721e7076d3062f2f2ebaf3f2bac8ac67be53962baf4eb97108ae"} Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.676532 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.677187 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.688773 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.696861 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 08:25:06 crc kubenswrapper[4945]: I1206 08:25:06.869688 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 06 08:25:07 crc kubenswrapper[4945]: W1206 08:25:07.178846 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod737bd8cb_ebc6_441f_9f6c_0f13e356232c.slice/crio-e805138908b82d277f7eee6f03449ef2bfe37ab893e657404b3d997e747ee980 WatchSource:0}: Error finding container e805138908b82d277f7eee6f03449ef2bfe37ab893e657404b3d997e747ee980: Status 404 returned error can't find the container with id e805138908b82d277f7eee6f03449ef2bfe37ab893e657404b3d997e747ee980 Dec 06 08:25:07 crc kubenswrapper[4945]: I1206 08:25:07.560819 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 08:25:07 crc kubenswrapper[4945]: I1206 08:25:07.562241 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 08:25:07 crc kubenswrapper[4945]: I1206 08:25:07.562954 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 08:25:07 crc kubenswrapper[4945]: I1206 08:25:07.643920 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9","Type":"ContainerStarted","Data":"49086b2fc48ee085a88aa89b3fff259ee5d2edd19e6774648eb88aaf543bf2ee"} Dec 06 08:25:07 crc kubenswrapper[4945]: I1206 08:25:07.646069 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"737bd8cb-ebc6-441f-9f6c-0f13e356232c","Type":"ContainerStarted","Data":"e805138908b82d277f7eee6f03449ef2bfe37ab893e657404b3d997e747ee980"} Dec 06 08:25:07 crc kubenswrapper[4945]: I1206 08:25:07.646531 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 08:25:07 crc kubenswrapper[4945]: I1206 08:25:07.648502 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 08:25:07 crc kubenswrapper[4945]: I1206 08:25:07.651913 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 08:25:07 crc kubenswrapper[4945]: I1206 08:25:07.753260 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.170928 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.310717 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-etc-machine-id\") pod \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.310778 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data-custom\") pod \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.310869 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-logs\") pod \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.310860 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" (UID: "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.310986 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-combined-ca-bundle\") pod \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.311037 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-scripts\") pod \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.311076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9f65\" (UniqueName: \"kubernetes.io/projected/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-kube-api-access-d9f65\") pod \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.311109 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data\") pod \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\" (UID: \"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab\") " Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.311432 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-logs" (OuterVolumeSpecName: "logs") pod "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" (UID: "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.311869 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.311893 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.317719 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-kube-api-access-d9f65" (OuterVolumeSpecName: "kube-api-access-d9f65") pod "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" (UID: "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab"). InnerVolumeSpecName "kube-api-access-d9f65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.317965 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" (UID: "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.327446 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-scripts" (OuterVolumeSpecName: "scripts") pod "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" (UID: "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.361505 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" (UID: "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.363369 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data" (OuterVolumeSpecName: "config-data") pod "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" (UID: "666b36c8-1753-4bf2-b2ed-6a8c3c4112ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.413436 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.413716 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.413819 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9f65\" (UniqueName: \"kubernetes.io/projected/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-kube-api-access-d9f65\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.413881 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.413966 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.658176 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"737bd8cb-ebc6-441f-9f6c-0f13e356232c","Type":"ContainerStarted","Data":"fd261f430685fe6ab112c9716917fa70a64d403a4c5e1ac0e6a1053113006597"} Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.658229 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"737bd8cb-ebc6-441f-9f6c-0f13e356232c","Type":"ContainerStarted","Data":"722b196e35bd4c83be1171ddcab1732d25a9ad5995a74e5d63aaaf5ad6117233"} Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.661346 4945 generic.go:334] "Generic (PLEG): container finished" podID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" containerID="83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a" exitCode=0 Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.661410 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab","Type":"ContainerDied","Data":"83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a"} Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.661438 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"666b36c8-1753-4bf2-b2ed-6a8c3c4112ab","Type":"ContainerDied","Data":"61f6c0dafb9b6cb843d164d96a7de89ba8013f1125c61d172c2a07ffa2df1387"} Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.661471 4945 scope.go:117] "RemoveContainer" containerID="83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.661613 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.669780 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a15a0964-44c0-4a17-9f3e-42a5ed41f4e9","Type":"ContainerStarted","Data":"ee588a4d9f1e8dc3f3c8b8b002e432459aeaf66dab9d07103d9fd5320dc59c84"} Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.699121 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.294300377 podStartE2EDuration="3.699103427s" podCreationTimestamp="2025-12-06 08:25:05 +0000 UTC" firstStartedPulling="2025-12-06 08:25:07.181800542 +0000 UTC m=+5520.636661586" lastFinishedPulling="2025-12-06 08:25:07.586603592 +0000 UTC m=+5521.041464636" observedRunningTime="2025-12-06 08:25:08.691019331 +0000 UTC m=+5522.145880405" watchObservedRunningTime="2025-12-06 08:25:08.699103427 +0000 UTC m=+5522.153964471" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.719573 4945 scope.go:117] "RemoveContainer" containerID="6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.724971 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.738853 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.746538 4945 scope.go:117] "RemoveContainer" containerID="83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a" Dec 06 08:25:08 crc kubenswrapper[4945]: E1206 08:25:08.753126 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a\": container with ID starting with 83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a not found: ID does not exist" containerID="83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.753200 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a"} err="failed to get container status \"83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a\": rpc error: code = NotFound desc = could not find container \"83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a\": container with ID starting with 83cf06affd6211bf12a6dc5982e5a16fde6c1615a6e03d0e73856ce346241e6a not found: ID does not exist" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.753234 4945 scope.go:117] "RemoveContainer" containerID="6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0" Dec 06 08:25:08 crc kubenswrapper[4945]: E1206 08:25:08.753533 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0\": container with ID starting with 6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0 not found: ID does not exist" containerID="6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.753562 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0"} err="failed to get container status \"6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0\": rpc error: code = NotFound desc = could not find container \"6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0\": container with ID starting with 6e67ac23881f636a159caceadd379a7a37296acfff112081329b0b05fd5204e0 not found: ID does not exist" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.756353 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 08:25:08 crc kubenswrapper[4945]: E1206 08:25:08.756794 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" containerName="cinder-api-log" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.756814 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" containerName="cinder-api-log" Dec 06 08:25:08 crc kubenswrapper[4945]: E1206 08:25:08.756851 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" containerName="cinder-api" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.756860 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" containerName="cinder-api" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.757083 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" containerName="cinder-api" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.757106 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" containerName="cinder-api-log" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.758146 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.760588 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.762322 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.6208806 podStartE2EDuration="3.762301668s" podCreationTimestamp="2025-12-06 08:25:05 +0000 UTC" firstStartedPulling="2025-12-06 08:25:06.085492161 +0000 UTC m=+5519.540353205" lastFinishedPulling="2025-12-06 08:25:07.226913229 +0000 UTC m=+5520.681774273" observedRunningTime="2025-12-06 08:25:08.747115862 +0000 UTC m=+5522.201976916" watchObservedRunningTime="2025-12-06 08:25:08.762301668 +0000 UTC m=+5522.217162712" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.785925 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.932886 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/117f8acf-2899-4353-ad75-8855c73fa61e-logs\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.932938 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-scripts\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.933026 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/117f8acf-2899-4353-ad75-8855c73fa61e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.933135 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgpmg\" (UniqueName: \"kubernetes.io/projected/117f8acf-2899-4353-ad75-8855c73fa61e-kube-api-access-vgpmg\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.933177 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.933212 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-config-data-custom\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.933365 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-config-data\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:08 crc kubenswrapper[4945]: I1206 08:25:08.964921 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="666b36c8-1753-4bf2-b2ed-6a8c3c4112ab" path="/var/lib/kubelet/pods/666b36c8-1753-4bf2-b2ed-6a8c3c4112ab/volumes" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.034926 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-config-data-custom\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.035017 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-config-data\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.035099 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/117f8acf-2899-4353-ad75-8855c73fa61e-logs\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.035130 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-scripts\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.035220 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/117f8acf-2899-4353-ad75-8855c73fa61e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.035379 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgpmg\" (UniqueName: \"kubernetes.io/projected/117f8acf-2899-4353-ad75-8855c73fa61e-kube-api-access-vgpmg\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.035423 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.035701 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/117f8acf-2899-4353-ad75-8855c73fa61e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.035890 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/117f8acf-2899-4353-ad75-8855c73fa61e-logs\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.042883 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-config-data-custom\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.043878 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.051462 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-scripts\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.054741 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117f8acf-2899-4353-ad75-8855c73fa61e-config-data\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.068315 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgpmg\" (UniqueName: \"kubernetes.io/projected/117f8acf-2899-4353-ad75-8855c73fa61e-kube-api-access-vgpmg\") pod \"cinder-api-0\" (UID: \"117f8acf-2899-4353-ad75-8855c73fa61e\") " pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.076918 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.578238 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 08:25:09 crc kubenswrapper[4945]: W1206 08:25:09.597768 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod117f8acf_2899_4353_ad75_8855c73fa61e.slice/crio-320494214c2550d1c44d781b43bcc248fd7a99f8c5dab496f5654affc3ef521a WatchSource:0}: Error finding container 320494214c2550d1c44d781b43bcc248fd7a99f8c5dab496f5654affc3ef521a: Status 404 returned error can't find the container with id 320494214c2550d1c44d781b43bcc248fd7a99f8c5dab496f5654affc3ef521a Dec 06 08:25:09 crc kubenswrapper[4945]: I1206 08:25:09.710040 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"117f8acf-2899-4353-ad75-8855c73fa61e","Type":"ContainerStarted","Data":"320494214c2550d1c44d781b43bcc248fd7a99f8c5dab496f5654affc3ef521a"} Dec 06 08:25:10 crc kubenswrapper[4945]: I1206 08:25:10.460817 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:11 crc kubenswrapper[4945]: I1206 08:25:11.299448 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 06 08:25:12 crc kubenswrapper[4945]: I1206 08:25:12.761723 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"117f8acf-2899-4353-ad75-8855c73fa61e","Type":"ContainerStarted","Data":"4a551f9f5fe27056c1defb3f5c12851c3056f92ae53c3e568916e246e4f9e169"} Dec 06 08:25:12 crc kubenswrapper[4945]: I1206 08:25:12.953250 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:25:12 crc kubenswrapper[4945]: E1206 08:25:12.953567 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:25:12 crc kubenswrapper[4945]: I1206 08:25:12.979789 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 08:25:13 crc kubenswrapper[4945]: I1206 08:25:13.040182 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 08:25:13 crc kubenswrapper[4945]: I1206 08:25:13.771603 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"117f8acf-2899-4353-ad75-8855c73fa61e","Type":"ContainerStarted","Data":"5648eb8efb23e43b8c898952256e3d5f3d886da243b43dfd0d69ebbd42e27c6b"} Dec 06 08:25:13 crc kubenswrapper[4945]: I1206 08:25:13.771732 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" containerName="cinder-scheduler" containerID="cri-o://10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88" gracePeriod=30 Dec 06 08:25:13 crc kubenswrapper[4945]: I1206 08:25:13.771847 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" containerName="probe" containerID="cri-o://7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db" gracePeriod=30 Dec 06 08:25:13 crc kubenswrapper[4945]: I1206 08:25:13.793388 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.793367322 podStartE2EDuration="5.793367322s" podCreationTimestamp="2025-12-06 08:25:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:25:13.79177886 +0000 UTC m=+5527.246639924" watchObservedRunningTime="2025-12-06 08:25:13.793367322 +0000 UTC m=+5527.248228366" Dec 06 08:25:14 crc kubenswrapper[4945]: I1206 08:25:14.079089 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 08:25:14 crc kubenswrapper[4945]: I1206 08:25:14.790362 4945 generic.go:334] "Generic (PLEG): container finished" podID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" containerID="10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88" exitCode=0 Dec 06 08:25:14 crc kubenswrapper[4945]: I1206 08:25:14.791371 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60","Type":"ContainerDied","Data":"10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88"} Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.144474 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.256688 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-scripts\") pod \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.256739 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bm2p\" (UniqueName: \"kubernetes.io/projected/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-kube-api-access-5bm2p\") pod \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.256763 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-combined-ca-bundle\") pod \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.256790 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data\") pod \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.256810 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-etc-machine-id\") pod \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.256860 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data-custom\") pod \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\" (UID: \"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60\") " Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.257648 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" (UID: "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.262788 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-scripts" (OuterVolumeSpecName: "scripts") pod "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" (UID: "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.264539 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-kube-api-access-5bm2p" (OuterVolumeSpecName: "kube-api-access-5bm2p") pod "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" (UID: "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60"). InnerVolumeSpecName "kube-api-access-5bm2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.275412 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" (UID: "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.320086 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" (UID: "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.359301 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.359333 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bm2p\" (UniqueName: \"kubernetes.io/projected/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-kube-api-access-5bm2p\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.359343 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.359353 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.359361 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.369643 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data" (OuterVolumeSpecName: "config-data") pod "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" (UID: "788bc4d9-95d2-4b50-bf1b-3d5c9f844f60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.462317 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.680997 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.801419 4945 generic.go:334] "Generic (PLEG): container finished" podID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" containerID="7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db" exitCode=0 Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.801479 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.801527 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60","Type":"ContainerDied","Data":"7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db"} Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.801560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"788bc4d9-95d2-4b50-bf1b-3d5c9f844f60","Type":"ContainerDied","Data":"7e98b0dc3d6e9251165b34ff57cec947d6cc04651718638999486fb97017d2cd"} Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.801582 4945 scope.go:117] "RemoveContainer" containerID="7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.842879 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.853766 4945 scope.go:117] "RemoveContainer" containerID="10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.863823 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.876778 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 08:25:15 crc kubenswrapper[4945]: E1206 08:25:15.877344 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" containerName="probe" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.877369 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" containerName="probe" Dec 06 08:25:15 crc kubenswrapper[4945]: E1206 08:25:15.877396 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" containerName="cinder-scheduler" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.877405 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" containerName="cinder-scheduler" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.877630 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" containerName="cinder-scheduler" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.877668 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" containerName="probe" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.878962 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.881434 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.891786 4945 scope.go:117] "RemoveContainer" containerID="7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.893034 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 08:25:15 crc kubenswrapper[4945]: E1206 08:25:15.921714 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db\": container with ID starting with 7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db not found: ID does not exist" containerID="7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.921769 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db"} err="failed to get container status \"7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db\": rpc error: code = NotFound desc = could not find container \"7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db\": container with ID starting with 7db296c8a0443a83e3603f70e519abc76753931f0a40b7a6b67a92aad43a92db not found: ID does not exist" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.921802 4945 scope.go:117] "RemoveContainer" containerID="10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88" Dec 06 08:25:15 crc kubenswrapper[4945]: E1206 08:25:15.922310 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88\": container with ID starting with 10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88 not found: ID does not exist" containerID="10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.922343 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88"} err="failed to get container status \"10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88\": rpc error: code = NotFound desc = could not find container \"10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88\": container with ID starting with 10815ce10fde8e386027502db74087253e440133a4d0c4f8cf8c95adff2d1f88 not found: ID does not exist" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.971869 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed728ae8-6b6b-4edd-b577-91ef23d729e9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.972006 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.972047 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvv8b\" (UniqueName: \"kubernetes.io/projected/ed728ae8-6b6b-4edd-b577-91ef23d729e9-kube-api-access-cvv8b\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.972224 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-config-data\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.972325 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-scripts\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:15 crc kubenswrapper[4945]: I1206 08:25:15.973614 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.075680 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-config-data\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.075990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-scripts\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.076089 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.076115 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed728ae8-6b6b-4edd-b577-91ef23d729e9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.076161 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.076181 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvv8b\" (UniqueName: \"kubernetes.io/projected/ed728ae8-6b6b-4edd-b577-91ef23d729e9-kube-api-access-cvv8b\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.076486 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed728ae8-6b6b-4edd-b577-91ef23d729e9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.081271 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.080438 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.093805 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-scripts\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.094161 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed728ae8-6b6b-4edd-b577-91ef23d729e9-config-data\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.104672 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvv8b\" (UniqueName: \"kubernetes.io/projected/ed728ae8-6b6b-4edd-b577-91ef23d729e9-kube-api-access-cvv8b\") pod \"cinder-scheduler-0\" (UID: \"ed728ae8-6b6b-4edd-b577-91ef23d729e9\") " pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.251174 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.515376 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.703384 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.814012 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ed728ae8-6b6b-4edd-b577-91ef23d729e9","Type":"ContainerStarted","Data":"c7238f326863c5a5c49f6d4d3248ccc524c6379727fe0458f88f42e3b85fb1bf"} Dec 06 08:25:16 crc kubenswrapper[4945]: I1206 08:25:16.971362 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788bc4d9-95d2-4b50-bf1b-3d5c9f844f60" path="/var/lib/kubelet/pods/788bc4d9-95d2-4b50-bf1b-3d5c9f844f60/volumes" Dec 06 08:25:17 crc kubenswrapper[4945]: I1206 08:25:17.826714 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ed728ae8-6b6b-4edd-b577-91ef23d729e9","Type":"ContainerStarted","Data":"2744afb46bac64386aad26285fc15fa47d0713cfd8b28b036a9acaec167fb825"} Dec 06 08:25:18 crc kubenswrapper[4945]: I1206 08:25:18.840192 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ed728ae8-6b6b-4edd-b577-91ef23d729e9","Type":"ContainerStarted","Data":"68893d3e09422f35ed35db9f69ebde863fd9daaf3a6cb091c671a2e9ff623cdb"} Dec 06 08:25:18 crc kubenswrapper[4945]: I1206 08:25:18.864434 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.864412797 podStartE2EDuration="3.864412797s" podCreationTimestamp="2025-12-06 08:25:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:25:18.858942111 +0000 UTC m=+5532.313803155" watchObservedRunningTime="2025-12-06 08:25:18.864412797 +0000 UTC m=+5532.319273841" Dec 06 08:25:20 crc kubenswrapper[4945]: I1206 08:25:20.995528 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 06 08:25:21 crc kubenswrapper[4945]: I1206 08:25:21.251966 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 08:25:25 crc kubenswrapper[4945]: I1206 08:25:25.258067 4945 scope.go:117] "RemoveContainer" containerID="ababc9b1e1f5da53bf58818c8a38eb635e05c168eacca390ea6da51a37d60f2a" Dec 06 08:25:25 crc kubenswrapper[4945]: I1206 08:25:25.282709 4945 scope.go:117] "RemoveContainer" containerID="a71b8fd3a1a3aec0166661f0b1a35ffdd9bfcf95696ecee1b42033448ddf9094" Dec 06 08:25:26 crc kubenswrapper[4945]: I1206 08:25:26.452088 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 08:25:26 crc kubenswrapper[4945]: I1206 08:25:26.962099 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:25:26 crc kubenswrapper[4945]: E1206 08:25:26.962418 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:25:37 crc kubenswrapper[4945]: I1206 08:25:37.953693 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:25:37 crc kubenswrapper[4945]: E1206 08:25:37.954464 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:25:49 crc kubenswrapper[4945]: I1206 08:25:49.953476 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:25:49 crc kubenswrapper[4945]: E1206 08:25:49.954357 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:25:55 crc kubenswrapper[4945]: I1206 08:25:55.056434 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-3f05-account-create-update-5hrms"] Dec 06 08:25:55 crc kubenswrapper[4945]: I1206 08:25:55.064220 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-87vk8"] Dec 06 08:25:55 crc kubenswrapper[4945]: I1206 08:25:55.073041 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-87vk8"] Dec 06 08:25:55 crc kubenswrapper[4945]: I1206 08:25:55.081249 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-3f05-account-create-update-5hrms"] Dec 06 08:25:56 crc kubenswrapper[4945]: I1206 08:25:56.964754 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b486576-3473-4507-babe-9cadb07f9d17" path="/var/lib/kubelet/pods/8b486576-3473-4507-babe-9cadb07f9d17/volumes" Dec 06 08:25:56 crc kubenswrapper[4945]: I1206 08:25:56.965905 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4835330-f95a-4b6b-af02-5e780b5e35be" path="/var/lib/kubelet/pods/e4835330-f95a-4b6b-af02-5e780b5e35be/volumes" Dec 06 08:26:00 crc kubenswrapper[4945]: I1206 08:26:00.953883 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:26:00 crc kubenswrapper[4945]: E1206 08:26:00.954552 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:26:07 crc kubenswrapper[4945]: I1206 08:26:07.033486 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-s85wb"] Dec 06 08:26:07 crc kubenswrapper[4945]: I1206 08:26:07.046376 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-s85wb"] Dec 06 08:26:08 crc kubenswrapper[4945]: I1206 08:26:08.965972 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79b89fa1-ca85-46d1-a9dd-0c15ef2052e3" path="/var/lib/kubelet/pods/79b89fa1-ca85-46d1-a9dd-0c15ef2052e3/volumes" Dec 06 08:26:13 crc kubenswrapper[4945]: I1206 08:26:13.953364 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:26:14 crc kubenswrapper[4945]: I1206 08:26:14.380641 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"170cc98b8d9a5021f285e8b2c4ca1f3a5177f12dc09aebb616fcafb0121da204"} Dec 06 08:26:20 crc kubenswrapper[4945]: I1206 08:26:20.042836 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-t6tq5"] Dec 06 08:26:20 crc kubenswrapper[4945]: I1206 08:26:20.050709 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-t6tq5"] Dec 06 08:26:20 crc kubenswrapper[4945]: I1206 08:26:20.962892 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c16c276f-ad77-45ab-b719-368517cf919d" path="/var/lib/kubelet/pods/c16c276f-ad77-45ab-b719-368517cf919d/volumes" Dec 06 08:26:25 crc kubenswrapper[4945]: I1206 08:26:25.513324 4945 scope.go:117] "RemoveContainer" containerID="e5bd8b31440c0e0732c0f41fe457000b4c885d5cd8e02288919d80fbb4a80801" Dec 06 08:26:25 crc kubenswrapper[4945]: I1206 08:26:25.561562 4945 scope.go:117] "RemoveContainer" containerID="1c0bc22c75aeb954f03758e5eb5a43e4189fe1d2a667b041b4ff21d91bee7eeb" Dec 06 08:26:25 crc kubenswrapper[4945]: I1206 08:26:25.601145 4945 scope.go:117] "RemoveContainer" containerID="6cb22d807b8bd6a4e00c6de4998d605c92b6fabfe0dc506cee6463e3ce354e41" Dec 06 08:26:25 crc kubenswrapper[4945]: I1206 08:26:25.630501 4945 scope.go:117] "RemoveContainer" containerID="5d7d7fff4ebe622d4e57e30a216d7afec2b897cc2b259598d1d2f4f77e39bbe0" Dec 06 08:26:59 crc kubenswrapper[4945]: I1206 08:26:59.977319 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-888dc8c77-7hkkf"] Dec 06 08:26:59 crc kubenswrapper[4945]: I1206 08:26:59.981234 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:26:59 crc kubenswrapper[4945]: I1206 08:26:59.987843 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 06 08:26:59 crc kubenswrapper[4945]: I1206 08:26:59.988157 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 06 08:26:59 crc kubenswrapper[4945]: I1206 08:26:59.988526 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-n9t9j" Dec 06 08:26:59 crc kubenswrapper[4945]: I1206 08:26:59.991962 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 06 08:26:59 crc kubenswrapper[4945]: I1206 08:26:59.997401 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-888dc8c77-7hkkf"] Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.049161 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.049420 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" containerName="glance-log" containerID="cri-o://2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42" gracePeriod=30 Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.049519 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" containerName="glance-httpd" containerID="cri-o://96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060" gracePeriod=30 Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.073125 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5e872eee-469e-480b-9eca-c89aa6a1be40-horizon-secret-key\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.073271 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e872eee-469e-480b-9eca-c89aa6a1be40-logs\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.073328 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t79n6\" (UniqueName: \"kubernetes.io/projected/5e872eee-469e-480b-9eca-c89aa6a1be40-kube-api-access-t79n6\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.073358 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-config-data\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.073385 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-scripts\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.106610 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-748744fc9c-m2ggc"] Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.108098 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.129513 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-748744fc9c-m2ggc"] Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.137600 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.137827 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" containerName="glance-log" containerID="cri-o://750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a" gracePeriod=30 Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.137912 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" containerName="glance-httpd" containerID="cri-o://7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155" gracePeriod=30 Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.175469 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e872eee-469e-480b-9eca-c89aa6a1be40-logs\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.175524 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t79n6\" (UniqueName: \"kubernetes.io/projected/5e872eee-469e-480b-9eca-c89aa6a1be40-kube-api-access-t79n6\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.175560 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-config-data\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.175586 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-scripts\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.175622 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5e872eee-469e-480b-9eca-c89aa6a1be40-horizon-secret-key\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.176648 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e872eee-469e-480b-9eca-c89aa6a1be40-logs\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.176672 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-scripts\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.178131 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-config-data\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.181578 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5e872eee-469e-480b-9eca-c89aa6a1be40-horizon-secret-key\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.197846 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t79n6\" (UniqueName: \"kubernetes.io/projected/5e872eee-469e-480b-9eca-c89aa6a1be40-kube-api-access-t79n6\") pod \"horizon-888dc8c77-7hkkf\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.277210 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-config-data\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.277266 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c41402cd-e31f-47ed-ba7a-1f352c1c819c-logs\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.277433 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c41402cd-e31f-47ed-ba7a-1f352c1c819c-horizon-secret-key\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.277474 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-scripts\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.277514 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpxv7\" (UniqueName: \"kubernetes.io/projected/c41402cd-e31f-47ed-ba7a-1f352c1c819c-kube-api-access-bpxv7\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.300486 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.378647 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-config-data\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.378684 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c41402cd-e31f-47ed-ba7a-1f352c1c819c-logs\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.378812 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c41402cd-e31f-47ed-ba7a-1f352c1c819c-horizon-secret-key\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.378847 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-scripts\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.378879 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpxv7\" (UniqueName: \"kubernetes.io/projected/c41402cd-e31f-47ed-ba7a-1f352c1c819c-kube-api-access-bpxv7\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.379723 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-scripts\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.379934 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c41402cd-e31f-47ed-ba7a-1f352c1c819c-logs\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.380199 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-config-data\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.382425 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c41402cd-e31f-47ed-ba7a-1f352c1c819c-horizon-secret-key\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.399806 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpxv7\" (UniqueName: \"kubernetes.io/projected/c41402cd-e31f-47ed-ba7a-1f352c1c819c-kube-api-access-bpxv7\") pod \"horizon-748744fc9c-m2ggc\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.430411 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.663106 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-888dc8c77-7hkkf"] Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.699618 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67bcf67b9f-fln2w"] Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.706455 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.734918 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67bcf67b9f-fln2w"] Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.779730 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-888dc8c77-7hkkf"] Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.787814 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.890348 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-scripts\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.890436 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7547017f-a981-4309-979f-6776230b2198-horizon-secret-key\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.890494 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5g5b\" (UniqueName: \"kubernetes.io/projected/7547017f-a981-4309-979f-6776230b2198-kube-api-access-h5g5b\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.890542 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7547017f-a981-4309-979f-6776230b2198-logs\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.890741 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-config-data\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.918000 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-748744fc9c-m2ggc"] Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.918442 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-888dc8c77-7hkkf" event={"ID":"5e872eee-469e-480b-9eca-c89aa6a1be40","Type":"ContainerStarted","Data":"a0eef28ffd74d29840e42909784df05dd47cfc916dd1367e6fdc05886c350e23"} Dec 06 08:27:00 crc kubenswrapper[4945]: W1206 08:27:00.920515 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc41402cd_e31f_47ed_ba7a_1f352c1c819c.slice/crio-efa44bdcc5772d20a808f848108f36123d05475eed5dfb2f7fb238318256b250 WatchSource:0}: Error finding container efa44bdcc5772d20a808f848108f36123d05475eed5dfb2f7fb238318256b250: Status 404 returned error can't find the container with id efa44bdcc5772d20a808f848108f36123d05475eed5dfb2f7fb238318256b250 Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.920708 4945 generic.go:334] "Generic (PLEG): container finished" podID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" containerID="750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a" exitCode=143 Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.920775 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6cb272d-4c37-4914-95e5-bcc1461b6ee9","Type":"ContainerDied","Data":"750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a"} Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.923873 4945 generic.go:334] "Generic (PLEG): container finished" podID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" containerID="2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42" exitCode=143 Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.923901 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c","Type":"ContainerDied","Data":"2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42"} Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.992683 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-config-data\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.992798 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-scripts\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.992830 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7547017f-a981-4309-979f-6776230b2198-horizon-secret-key\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.992867 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5g5b\" (UniqueName: \"kubernetes.io/projected/7547017f-a981-4309-979f-6776230b2198-kube-api-access-h5g5b\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.992897 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7547017f-a981-4309-979f-6776230b2198-logs\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.993588 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7547017f-a981-4309-979f-6776230b2198-logs\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.994386 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-config-data\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.995453 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-scripts\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:00 crc kubenswrapper[4945]: I1206 08:27:00.999611 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7547017f-a981-4309-979f-6776230b2198-horizon-secret-key\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:01 crc kubenswrapper[4945]: I1206 08:27:01.009349 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5g5b\" (UniqueName: \"kubernetes.io/projected/7547017f-a981-4309-979f-6776230b2198-kube-api-access-h5g5b\") pod \"horizon-67bcf67b9f-fln2w\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:01 crc kubenswrapper[4945]: I1206 08:27:01.063951 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:01 crc kubenswrapper[4945]: W1206 08:27:01.522877 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7547017f_a981_4309_979f_6776230b2198.slice/crio-2cd687e4863598523474dcc1ed986c4a6e41f7b317f5ffa4835bb8675d2ae822 WatchSource:0}: Error finding container 2cd687e4863598523474dcc1ed986c4a6e41f7b317f5ffa4835bb8675d2ae822: Status 404 returned error can't find the container with id 2cd687e4863598523474dcc1ed986c4a6e41f7b317f5ffa4835bb8675d2ae822 Dec 06 08:27:01 crc kubenswrapper[4945]: I1206 08:27:01.525857 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67bcf67b9f-fln2w"] Dec 06 08:27:01 crc kubenswrapper[4945]: I1206 08:27:01.936342 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748744fc9c-m2ggc" event={"ID":"c41402cd-e31f-47ed-ba7a-1f352c1c819c","Type":"ContainerStarted","Data":"efa44bdcc5772d20a808f848108f36123d05475eed5dfb2f7fb238318256b250"} Dec 06 08:27:01 crc kubenswrapper[4945]: I1206 08:27:01.939582 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bcf67b9f-fln2w" event={"ID":"7547017f-a981-4309-979f-6776230b2198","Type":"ContainerStarted","Data":"2cd687e4863598523474dcc1ed986c4a6e41f7b317f5ffa4835bb8675d2ae822"} Dec 06 08:27:03 crc kubenswrapper[4945]: E1206 08:27:03.404948 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd6770a4_0afe_4cd1_bab2_6cb846c8d09c.slice/crio-96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd6770a4_0afe_4cd1_bab2_6cb846c8d09c.slice/crio-conmon-96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060.scope\": RecentStats: unable to find data in memory cache]" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.737714 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.796809 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.862927 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-config-data\") pod \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.863025 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-httpd-run\") pod \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.863069 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-logs\") pod \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.863092 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-combined-ca-bundle\") pod \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.863184 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-ceph\") pod \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.863237 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-scripts\") pod \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.863307 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx6kv\" (UniqueName: \"kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-kube-api-access-dx6kv\") pod \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\" (UID: \"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.865390 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-logs" (OuterVolumeSpecName: "logs") pod "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" (UID: "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.866977 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" (UID: "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.872462 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-scripts" (OuterVolumeSpecName: "scripts") pod "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" (UID: "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.873102 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-kube-api-access-dx6kv" (OuterVolumeSpecName: "kube-api-access-dx6kv") pod "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" (UID: "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c"). InnerVolumeSpecName "kube-api-access-dx6kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.876018 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-ceph" (OuterVolumeSpecName: "ceph") pod "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" (UID: "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.912719 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" (UID: "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.929661 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-config-data" (OuterVolumeSpecName: "config-data") pod "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" (UID: "dd6770a4-0afe-4cd1-bab2-6cb846c8d09c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.964424 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-config-data\") pod \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.964491 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-scripts\") pod \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.964515 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rphn\" (UniqueName: \"kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-kube-api-access-8rphn\") pod \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.964539 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-httpd-run\") pod \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.964688 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-logs\") pod \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.964718 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-ceph\") pod \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.964740 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-combined-ca-bundle\") pod \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\" (UID: \"a6cb272d-4c37-4914-95e5-bcc1461b6ee9\") " Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.965088 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.965105 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.965115 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx6kv\" (UniqueName: \"kubernetes.io/projected/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-kube-api-access-dx6kv\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.965124 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.965133 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.965141 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.965149 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.965540 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a6cb272d-4c37-4914-95e5-bcc1461b6ee9" (UID: "a6cb272d-4c37-4914-95e5-bcc1461b6ee9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.966004 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-logs" (OuterVolumeSpecName: "logs") pod "a6cb272d-4c37-4914-95e5-bcc1461b6ee9" (UID: "a6cb272d-4c37-4914-95e5-bcc1461b6ee9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.967626 4945 generic.go:334] "Generic (PLEG): container finished" podID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" containerID="96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060" exitCode=0 Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.967708 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.967734 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c","Type":"ContainerDied","Data":"96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060"} Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.967772 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dd6770a4-0afe-4cd1-bab2-6cb846c8d09c","Type":"ContainerDied","Data":"35e98d7c50d2924403e245020eaff16deb69bef23bdf61abfd2f4d79bf289744"} Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.967795 4945 scope.go:117] "RemoveContainer" containerID="96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.968143 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-scripts" (OuterVolumeSpecName: "scripts") pod "a6cb272d-4c37-4914-95e5-bcc1461b6ee9" (UID: "a6cb272d-4c37-4914-95e5-bcc1461b6ee9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.969667 4945 generic.go:334] "Generic (PLEG): container finished" podID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" containerID="7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155" exitCode=0 Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.969688 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6cb272d-4c37-4914-95e5-bcc1461b6ee9","Type":"ContainerDied","Data":"7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155"} Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.969702 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6cb272d-4c37-4914-95e5-bcc1461b6ee9","Type":"ContainerDied","Data":"1bcff3830c59f5fef73ce7393f627bd603ae76a00c371d34e6a4561d9d5cbd87"} Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.969759 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.972069 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-ceph" (OuterVolumeSpecName: "ceph") pod "a6cb272d-4c37-4914-95e5-bcc1461b6ee9" (UID: "a6cb272d-4c37-4914-95e5-bcc1461b6ee9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.974534 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-kube-api-access-8rphn" (OuterVolumeSpecName: "kube-api-access-8rphn") pod "a6cb272d-4c37-4914-95e5-bcc1461b6ee9" (UID: "a6cb272d-4c37-4914-95e5-bcc1461b6ee9"). InnerVolumeSpecName "kube-api-access-8rphn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:27:03 crc kubenswrapper[4945]: I1206 08:27:03.999446 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6cb272d-4c37-4914-95e5-bcc1461b6ee9" (UID: "a6cb272d-4c37-4914-95e5-bcc1461b6ee9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.000230 4945 scope.go:117] "RemoveContainer" containerID="2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.002974 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.015953 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.028161 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:27:04 crc kubenswrapper[4945]: E1206 08:27:04.028691 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" containerName="glance-log" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.028708 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" containerName="glance-log" Dec 06 08:27:04 crc kubenswrapper[4945]: E1206 08:27:04.028732 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" containerName="glance-log" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.028740 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" containerName="glance-log" Dec 06 08:27:04 crc kubenswrapper[4945]: E1206 08:27:04.028749 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" containerName="glance-httpd" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.028756 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" containerName="glance-httpd" Dec 06 08:27:04 crc kubenswrapper[4945]: E1206 08:27:04.028772 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" containerName="glance-httpd" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.028781 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" containerName="glance-httpd" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.029002 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" containerName="glance-httpd" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.029018 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" containerName="glance-log" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.029033 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" containerName="glance-httpd" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.029056 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" containerName="glance-log" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.032184 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.032374 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-config-data" (OuterVolumeSpecName: "config-data") pod "a6cb272d-4c37-4914-95e5-bcc1461b6ee9" (UID: "a6cb272d-4c37-4914-95e5-bcc1461b6ee9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.035840 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.042331 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.060669 4945 scope.go:117] "RemoveContainer" containerID="96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060" Dec 06 08:27:04 crc kubenswrapper[4945]: E1206 08:27:04.061705 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060\": container with ID starting with 96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060 not found: ID does not exist" containerID="96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.061750 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060"} err="failed to get container status \"96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060\": rpc error: code = NotFound desc = could not find container \"96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060\": container with ID starting with 96aeb2b8b81b152a2a53d9a9d45fd16874c24900dede1185997105e5d74d1060 not found: ID does not exist" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.061775 4945 scope.go:117] "RemoveContainer" containerID="2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42" Dec 06 08:27:04 crc kubenswrapper[4945]: E1206 08:27:04.062659 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42\": container with ID starting with 2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42 not found: ID does not exist" containerID="2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.062697 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42"} err="failed to get container status \"2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42\": rpc error: code = NotFound desc = could not find container \"2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42\": container with ID starting with 2d8c8c5d9d63908e5954939a5f21501fb925b11bafe48c0a5fa1323c19975e42 not found: ID does not exist" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.062725 4945 scope.go:117] "RemoveContainer" containerID="7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.068322 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.068345 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.068355 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.068363 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.068372 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rphn\" (UniqueName: \"kubernetes.io/projected/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-kube-api-access-8rphn\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.068381 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.068388 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6cb272d-4c37-4914-95e5-bcc1461b6ee9-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.088382 4945 scope.go:117] "RemoveContainer" containerID="750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.106670 4945 scope.go:117] "RemoveContainer" containerID="7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155" Dec 06 08:27:04 crc kubenswrapper[4945]: E1206 08:27:04.106979 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155\": container with ID starting with 7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155 not found: ID does not exist" containerID="7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.107011 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155"} err="failed to get container status \"7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155\": rpc error: code = NotFound desc = could not find container \"7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155\": container with ID starting with 7b80a9b86d7c1fbfb30e4d6c252938063e376984a81f7361c293ff4d75b96155 not found: ID does not exist" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.107034 4945 scope.go:117] "RemoveContainer" containerID="750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a" Dec 06 08:27:04 crc kubenswrapper[4945]: E1206 08:27:04.107236 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a\": container with ID starting with 750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a not found: ID does not exist" containerID="750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.107263 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a"} err="failed to get container status \"750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a\": rpc error: code = NotFound desc = could not find container \"750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a\": container with ID starting with 750ecc302f85b76e181a9dade48241f5f907715da71b5dae99a00b6853b7040a not found: ID does not exist" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.170905 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f2fce06-f02b-4812-a91d-be28a5b74705-logs\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.171035 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1f2fce06-f02b-4812-a91d-be28a5b74705-ceph\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.171151 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f2fce06-f02b-4812-a91d-be28a5b74705-scripts\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.171192 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hh6c\" (UniqueName: \"kubernetes.io/projected/1f2fce06-f02b-4812-a91d-be28a5b74705-kube-api-access-7hh6c\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.171505 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f2fce06-f02b-4812-a91d-be28a5b74705-config-data\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.171575 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f2fce06-f02b-4812-a91d-be28a5b74705-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.171630 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2fce06-f02b-4812-a91d-be28a5b74705-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.273668 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1f2fce06-f02b-4812-a91d-be28a5b74705-ceph\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.273751 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f2fce06-f02b-4812-a91d-be28a5b74705-scripts\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.273778 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hh6c\" (UniqueName: \"kubernetes.io/projected/1f2fce06-f02b-4812-a91d-be28a5b74705-kube-api-access-7hh6c\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.273883 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f2fce06-f02b-4812-a91d-be28a5b74705-config-data\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.273905 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f2fce06-f02b-4812-a91d-be28a5b74705-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.273939 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2fce06-f02b-4812-a91d-be28a5b74705-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.273978 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f2fce06-f02b-4812-a91d-be28a5b74705-logs\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.274660 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f2fce06-f02b-4812-a91d-be28a5b74705-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.274956 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f2fce06-f02b-4812-a91d-be28a5b74705-logs\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.278952 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f2fce06-f02b-4812-a91d-be28a5b74705-config-data\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.279475 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f2fce06-f02b-4812-a91d-be28a5b74705-scripts\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.279551 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2fce06-f02b-4812-a91d-be28a5b74705-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.280159 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1f2fce06-f02b-4812-a91d-be28a5b74705-ceph\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.304844 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hh6c\" (UniqueName: \"kubernetes.io/projected/1f2fce06-f02b-4812-a91d-be28a5b74705-kube-api-access-7hh6c\") pod \"glance-default-external-api-0\" (UID: \"1f2fce06-f02b-4812-a91d-be28a5b74705\") " pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.309127 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.337378 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.356402 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.356936 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.358446 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.361160 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.367571 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.478216 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r87fp\" (UniqueName: \"kubernetes.io/projected/84ceb1ce-0b26-4b51-bc94-de365bc9849b-kube-api-access-r87fp\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.478378 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ceb1ce-0b26-4b51-bc94-de365bc9849b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.478440 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84ceb1ce-0b26-4b51-bc94-de365bc9849b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.478477 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ceb1ce-0b26-4b51-bc94-de365bc9849b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.478499 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84ceb1ce-0b26-4b51-bc94-de365bc9849b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.478528 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84ceb1ce-0b26-4b51-bc94-de365bc9849b-logs\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.478612 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/84ceb1ce-0b26-4b51-bc94-de365bc9849b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.580807 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84ceb1ce-0b26-4b51-bc94-de365bc9849b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.580861 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ceb1ce-0b26-4b51-bc94-de365bc9849b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.580931 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84ceb1ce-0b26-4b51-bc94-de365bc9849b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.581032 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84ceb1ce-0b26-4b51-bc94-de365bc9849b-logs\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.581102 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/84ceb1ce-0b26-4b51-bc94-de365bc9849b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.581136 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r87fp\" (UniqueName: \"kubernetes.io/projected/84ceb1ce-0b26-4b51-bc94-de365bc9849b-kube-api-access-r87fp\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.581237 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ceb1ce-0b26-4b51-bc94-de365bc9849b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.582003 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84ceb1ce-0b26-4b51-bc94-de365bc9849b-logs\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.582408 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84ceb1ce-0b26-4b51-bc94-de365bc9849b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.586435 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84ceb1ce-0b26-4b51-bc94-de365bc9849b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.586570 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ceb1ce-0b26-4b51-bc94-de365bc9849b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.587140 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ceb1ce-0b26-4b51-bc94-de365bc9849b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.600940 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/84ceb1ce-0b26-4b51-bc94-de365bc9849b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.602159 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r87fp\" (UniqueName: \"kubernetes.io/projected/84ceb1ce-0b26-4b51-bc94-de365bc9849b-kube-api-access-r87fp\") pod \"glance-default-internal-api-0\" (UID: \"84ceb1ce-0b26-4b51-bc94-de365bc9849b\") " pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.721068 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.963558 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6cb272d-4c37-4914-95e5-bcc1461b6ee9" path="/var/lib/kubelet/pods/a6cb272d-4c37-4914-95e5-bcc1461b6ee9/volumes" Dec 06 08:27:04 crc kubenswrapper[4945]: I1206 08:27:04.964246 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd6770a4-0afe-4cd1-bab2-6cb846c8d09c" path="/var/lib/kubelet/pods/dd6770a4-0afe-4cd1-bab2-6cb846c8d09c/volumes" Dec 06 08:27:09 crc kubenswrapper[4945]: I1206 08:27:09.031214 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748744fc9c-m2ggc" event={"ID":"c41402cd-e31f-47ed-ba7a-1f352c1c819c","Type":"ContainerStarted","Data":"91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c"} Dec 06 08:27:09 crc kubenswrapper[4945]: I1206 08:27:09.042154 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bcf67b9f-fln2w" event={"ID":"7547017f-a981-4309-979f-6776230b2198","Type":"ContainerStarted","Data":"e490063ebf7044297abdd8d76cff3a7686ad4ed811ca3a265e5ef91372cfc76b"} Dec 06 08:27:09 crc kubenswrapper[4945]: I1206 08:27:09.046101 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-888dc8c77-7hkkf" event={"ID":"5e872eee-469e-480b-9eca-c89aa6a1be40","Type":"ContainerStarted","Data":"e04f7df339b627ce0a17c30cf9ac7646b0783c84a84ce512ade0da3a1fe57725"} Dec 06 08:27:09 crc kubenswrapper[4945]: W1206 08:27:09.182065 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f2fce06_f02b_4812_a91d_be28a5b74705.slice/crio-724088e4c8ef99bbf759dcb7213c36776eab5f8e1fba09ce95e076455b20b7ae WatchSource:0}: Error finding container 724088e4c8ef99bbf759dcb7213c36776eab5f8e1fba09ce95e076455b20b7ae: Status 404 returned error can't find the container with id 724088e4c8ef99bbf759dcb7213c36776eab5f8e1fba09ce95e076455b20b7ae Dec 06 08:27:09 crc kubenswrapper[4945]: I1206 08:27:09.188401 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 08:27:09 crc kubenswrapper[4945]: I1206 08:27:09.281958 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.055311 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1f2fce06-f02b-4812-a91d-be28a5b74705","Type":"ContainerStarted","Data":"6a9625b8d71e6903c304198eb587b914813ba41ce3713a7f214d5612a2e1ada7"} Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.056551 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1f2fce06-f02b-4812-a91d-be28a5b74705","Type":"ContainerStarted","Data":"724088e4c8ef99bbf759dcb7213c36776eab5f8e1fba09ce95e076455b20b7ae"} Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.057208 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748744fc9c-m2ggc" event={"ID":"c41402cd-e31f-47ed-ba7a-1f352c1c819c","Type":"ContainerStarted","Data":"828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d"} Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.065730 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bcf67b9f-fln2w" event={"ID":"7547017f-a981-4309-979f-6776230b2198","Type":"ContainerStarted","Data":"73f4abd92b9c37ed61a7ad565c7e25a5ca8dd2322dc464962eafe564a1fd020e"} Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.069097 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-888dc8c77-7hkkf" event={"ID":"5e872eee-469e-480b-9eca-c89aa6a1be40","Type":"ContainerStarted","Data":"4de9d8a13d81c33f94fe5eb59e73d977f3d7845d4c465d6acbff19b157ac21e3"} Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.069170 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-888dc8c77-7hkkf" podUID="5e872eee-469e-480b-9eca-c89aa6a1be40" containerName="horizon-log" containerID="cri-o://e04f7df339b627ce0a17c30cf9ac7646b0783c84a84ce512ade0da3a1fe57725" gracePeriod=30 Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.069178 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-888dc8c77-7hkkf" podUID="5e872eee-469e-480b-9eca-c89aa6a1be40" containerName="horizon" containerID="cri-o://4de9d8a13d81c33f94fe5eb59e73d977f3d7845d4c465d6acbff19b157ac21e3" gracePeriod=30 Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.082749 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84ceb1ce-0b26-4b51-bc94-de365bc9849b","Type":"ContainerStarted","Data":"6435b01a57264d81a5daa2c5e4270043637a6b013e06d2f9172b79769a56599b"} Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.082840 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84ceb1ce-0b26-4b51-bc94-de365bc9849b","Type":"ContainerStarted","Data":"d69d1d2f44a98a8de5a64899477fd8b27c877722fb7f028625171557823fc26f"} Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.087058 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-748744fc9c-m2ggc" podStartSLOduration=2.359546422 podStartE2EDuration="10.087036829s" podCreationTimestamp="2025-12-06 08:27:00 +0000 UTC" firstStartedPulling="2025-12-06 08:27:00.922979056 +0000 UTC m=+5634.377840100" lastFinishedPulling="2025-12-06 08:27:08.650469463 +0000 UTC m=+5642.105330507" observedRunningTime="2025-12-06 08:27:10.076101146 +0000 UTC m=+5643.530962200" watchObservedRunningTime="2025-12-06 08:27:10.087036829 +0000 UTC m=+5643.541897873" Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.119822 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-888dc8c77-7hkkf" podStartSLOduration=3.256884895 podStartE2EDuration="11.119801005s" podCreationTimestamp="2025-12-06 08:26:59 +0000 UTC" firstStartedPulling="2025-12-06 08:27:00.787551483 +0000 UTC m=+5634.242412527" lastFinishedPulling="2025-12-06 08:27:08.650467593 +0000 UTC m=+5642.105328637" observedRunningTime="2025-12-06 08:27:10.093580164 +0000 UTC m=+5643.548441228" watchObservedRunningTime="2025-12-06 08:27:10.119801005 +0000 UTC m=+5643.574662049" Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.127161 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67bcf67b9f-fln2w" podStartSLOduration=2.953407971 podStartE2EDuration="10.126216117s" podCreationTimestamp="2025-12-06 08:27:00 +0000 UTC" firstStartedPulling="2025-12-06 08:27:01.525390574 +0000 UTC m=+5634.980251618" lastFinishedPulling="2025-12-06 08:27:08.69819872 +0000 UTC m=+5642.153059764" observedRunningTime="2025-12-06 08:27:10.11285689 +0000 UTC m=+5643.567717954" watchObservedRunningTime="2025-12-06 08:27:10.126216117 +0000 UTC m=+5643.581077161" Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.301036 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.430964 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:10 crc kubenswrapper[4945]: I1206 08:27:10.431265 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:11 crc kubenswrapper[4945]: I1206 08:27:11.065108 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:11 crc kubenswrapper[4945]: I1206 08:27:11.065254 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:11 crc kubenswrapper[4945]: I1206 08:27:11.111797 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1f2fce06-f02b-4812-a91d-be28a5b74705","Type":"ContainerStarted","Data":"67b73d057760c0fd85e74fb42b20c60f74cb9949b028f88edac9072f98066fdc"} Dec 06 08:27:11 crc kubenswrapper[4945]: I1206 08:27:11.116101 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84ceb1ce-0b26-4b51-bc94-de365bc9849b","Type":"ContainerStarted","Data":"139cc2e0778ac34f68210e3a0f1118efaa156febd731547f22dc848dfaa75f5a"} Dec 06 08:27:11 crc kubenswrapper[4945]: I1206 08:27:11.153165 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.153145643 podStartE2EDuration="8.153145643s" podCreationTimestamp="2025-12-06 08:27:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:27:11.135896162 +0000 UTC m=+5644.590757216" watchObservedRunningTime="2025-12-06 08:27:11.153145643 +0000 UTC m=+5644.608006687" Dec 06 08:27:11 crc kubenswrapper[4945]: I1206 08:27:11.162346 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.162330719 podStartE2EDuration="7.162330719s" podCreationTimestamp="2025-12-06 08:27:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:27:11.159859713 +0000 UTC m=+5644.614720777" watchObservedRunningTime="2025-12-06 08:27:11.162330719 +0000 UTC m=+5644.617191763" Dec 06 08:27:14 crc kubenswrapper[4945]: I1206 08:27:14.357963 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 08:27:14 crc kubenswrapper[4945]: I1206 08:27:14.360051 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 08:27:14 crc kubenswrapper[4945]: I1206 08:27:14.389111 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 08:27:14 crc kubenswrapper[4945]: I1206 08:27:14.439908 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 08:27:14 crc kubenswrapper[4945]: I1206 08:27:14.722181 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:14 crc kubenswrapper[4945]: I1206 08:27:14.722236 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:14 crc kubenswrapper[4945]: I1206 08:27:14.758940 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:14 crc kubenswrapper[4945]: I1206 08:27:14.771117 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:15 crc kubenswrapper[4945]: I1206 08:27:15.180551 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:15 crc kubenswrapper[4945]: I1206 08:27:15.180599 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:15 crc kubenswrapper[4945]: I1206 08:27:15.181694 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 08:27:15 crc kubenswrapper[4945]: I1206 08:27:15.181719 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 08:27:18 crc kubenswrapper[4945]: I1206 08:27:18.244867 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 08:27:18 crc kubenswrapper[4945]: I1206 08:27:18.245986 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 08:27:18 crc kubenswrapper[4945]: I1206 08:27:18.248015 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 08:27:18 crc kubenswrapper[4945]: I1206 08:27:18.267611 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:18 crc kubenswrapper[4945]: I1206 08:27:18.271186 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 08:27:18 crc kubenswrapper[4945]: I1206 08:27:18.331626 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 08:27:20 crc kubenswrapper[4945]: I1206 08:27:20.433103 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-748744fc9c-m2ggc" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.91:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.91:8080: connect: connection refused" Dec 06 08:27:21 crc kubenswrapper[4945]: I1206 08:27:21.065811 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-67bcf67b9f-fln2w" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.92:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.92:8080: connect: connection refused" Dec 06 08:27:32 crc kubenswrapper[4945]: I1206 08:27:32.255557 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:32 crc kubenswrapper[4945]: I1206 08:27:32.805290 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:33 crc kubenswrapper[4945]: I1206 08:27:33.882893 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:27:34 crc kubenswrapper[4945]: I1206 08:27:34.580920 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:27:34 crc kubenswrapper[4945]: I1206 08:27:34.650236 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-748744fc9c-m2ggc"] Dec 06 08:27:34 crc kubenswrapper[4945]: I1206 08:27:34.650508 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-748744fc9c-m2ggc" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon-log" containerID="cri-o://91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c" gracePeriod=30 Dec 06 08:27:34 crc kubenswrapper[4945]: I1206 08:27:34.650563 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-748744fc9c-m2ggc" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon" containerID="cri-o://828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d" gracePeriod=30 Dec 06 08:27:38 crc kubenswrapper[4945]: I1206 08:27:38.390178 4945 generic.go:334] "Generic (PLEG): container finished" podID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerID="828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d" exitCode=0 Dec 06 08:27:38 crc kubenswrapper[4945]: I1206 08:27:38.390370 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748744fc9c-m2ggc" event={"ID":"c41402cd-e31f-47ed-ba7a-1f352c1c819c","Type":"ContainerDied","Data":"828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d"} Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.410621 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e872eee-469e-480b-9eca-c89aa6a1be40" containerID="4de9d8a13d81c33f94fe5eb59e73d977f3d7845d4c465d6acbff19b157ac21e3" exitCode=137 Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.410956 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e872eee-469e-480b-9eca-c89aa6a1be40" containerID="e04f7df339b627ce0a17c30cf9ac7646b0783c84a84ce512ade0da3a1fe57725" exitCode=137 Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.410813 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-888dc8c77-7hkkf" event={"ID":"5e872eee-469e-480b-9eca-c89aa6a1be40","Type":"ContainerDied","Data":"4de9d8a13d81c33f94fe5eb59e73d977f3d7845d4c465d6acbff19b157ac21e3"} Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.410989 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-888dc8c77-7hkkf" event={"ID":"5e872eee-469e-480b-9eca-c89aa6a1be40","Type":"ContainerDied","Data":"e04f7df339b627ce0a17c30cf9ac7646b0783c84a84ce512ade0da3a1fe57725"} Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.431490 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-748744fc9c-m2ggc" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.91:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.91:8080: connect: connection refused" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.561325 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.693158 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e872eee-469e-480b-9eca-c89aa6a1be40-logs\") pod \"5e872eee-469e-480b-9eca-c89aa6a1be40\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.693211 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-config-data\") pod \"5e872eee-469e-480b-9eca-c89aa6a1be40\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.693707 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e872eee-469e-480b-9eca-c89aa6a1be40-logs" (OuterVolumeSpecName: "logs") pod "5e872eee-469e-480b-9eca-c89aa6a1be40" (UID: "5e872eee-469e-480b-9eca-c89aa6a1be40"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.693272 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t79n6\" (UniqueName: \"kubernetes.io/projected/5e872eee-469e-480b-9eca-c89aa6a1be40-kube-api-access-t79n6\") pod \"5e872eee-469e-480b-9eca-c89aa6a1be40\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.694239 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5e872eee-469e-480b-9eca-c89aa6a1be40-horizon-secret-key\") pod \"5e872eee-469e-480b-9eca-c89aa6a1be40\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.694373 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-scripts\") pod \"5e872eee-469e-480b-9eca-c89aa6a1be40\" (UID: \"5e872eee-469e-480b-9eca-c89aa6a1be40\") " Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.694751 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e872eee-469e-480b-9eca-c89aa6a1be40-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.698824 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e872eee-469e-480b-9eca-c89aa6a1be40-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5e872eee-469e-480b-9eca-c89aa6a1be40" (UID: "5e872eee-469e-480b-9eca-c89aa6a1be40"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.698842 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e872eee-469e-480b-9eca-c89aa6a1be40-kube-api-access-t79n6" (OuterVolumeSpecName: "kube-api-access-t79n6") pod "5e872eee-469e-480b-9eca-c89aa6a1be40" (UID: "5e872eee-469e-480b-9eca-c89aa6a1be40"). InnerVolumeSpecName "kube-api-access-t79n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.716539 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-config-data" (OuterVolumeSpecName: "config-data") pod "5e872eee-469e-480b-9eca-c89aa6a1be40" (UID: "5e872eee-469e-480b-9eca-c89aa6a1be40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.716768 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-scripts" (OuterVolumeSpecName: "scripts") pod "5e872eee-469e-480b-9eca-c89aa6a1be40" (UID: "5e872eee-469e-480b-9eca-c89aa6a1be40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.796570 4945 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5e872eee-469e-480b-9eca-c89aa6a1be40-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.796607 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.796616 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e872eee-469e-480b-9eca-c89aa6a1be40-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:40 crc kubenswrapper[4945]: I1206 08:27:40.796625 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t79n6\" (UniqueName: \"kubernetes.io/projected/5e872eee-469e-480b-9eca-c89aa6a1be40-kube-api-access-t79n6\") on node \"crc\" DevicePath \"\"" Dec 06 08:27:41 crc kubenswrapper[4945]: I1206 08:27:41.420142 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-888dc8c77-7hkkf" event={"ID":"5e872eee-469e-480b-9eca-c89aa6a1be40","Type":"ContainerDied","Data":"a0eef28ffd74d29840e42909784df05dd47cfc916dd1367e6fdc05886c350e23"} Dec 06 08:27:41 crc kubenswrapper[4945]: I1206 08:27:41.420209 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-888dc8c77-7hkkf" Dec 06 08:27:41 crc kubenswrapper[4945]: I1206 08:27:41.420219 4945 scope.go:117] "RemoveContainer" containerID="4de9d8a13d81c33f94fe5eb59e73d977f3d7845d4c465d6acbff19b157ac21e3" Dec 06 08:27:41 crc kubenswrapper[4945]: I1206 08:27:41.444298 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-888dc8c77-7hkkf"] Dec 06 08:27:41 crc kubenswrapper[4945]: I1206 08:27:41.452357 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-888dc8c77-7hkkf"] Dec 06 08:27:41 crc kubenswrapper[4945]: I1206 08:27:41.582683 4945 scope.go:117] "RemoveContainer" containerID="e04f7df339b627ce0a17c30cf9ac7646b0783c84a84ce512ade0da3a1fe57725" Dec 06 08:27:42 crc kubenswrapper[4945]: I1206 08:27:42.967814 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e872eee-469e-480b-9eca-c89aa6a1be40" path="/var/lib/kubelet/pods/5e872eee-469e-480b-9eca-c89aa6a1be40/volumes" Dec 06 08:27:50 crc kubenswrapper[4945]: I1206 08:27:50.430975 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-748744fc9c-m2ggc" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.91:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.91:8080: connect: connection refused" Dec 06 08:28:00 crc kubenswrapper[4945]: I1206 08:28:00.431733 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-748744fc9c-m2ggc" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.91:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.91:8080: connect: connection refused" Dec 06 08:28:00 crc kubenswrapper[4945]: I1206 08:28:00.432502 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:28:04 crc kubenswrapper[4945]: E1206 08:28:04.967373 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc41402cd_e31f_47ed_ba7a_1f352c1c819c.slice/crio-conmon-91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c.scope\": RecentStats: unable to find data in memory cache]" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.545673 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.686239 4945 generic.go:334] "Generic (PLEG): container finished" podID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerID="91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c" exitCode=137 Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.686304 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748744fc9c-m2ggc" event={"ID":"c41402cd-e31f-47ed-ba7a-1f352c1c819c","Type":"ContainerDied","Data":"91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c"} Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.686350 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748744fc9c-m2ggc" event={"ID":"c41402cd-e31f-47ed-ba7a-1f352c1c819c","Type":"ContainerDied","Data":"efa44bdcc5772d20a808f848108f36123d05475eed5dfb2f7fb238318256b250"} Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.686373 4945 scope.go:117] "RemoveContainer" containerID="828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.686416 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-scripts\") pod \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.686369 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-748744fc9c-m2ggc" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.686563 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c41402cd-e31f-47ed-ba7a-1f352c1c819c-horizon-secret-key\") pod \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.688166 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-config-data\") pod \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.689037 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c41402cd-e31f-47ed-ba7a-1f352c1c819c-logs\") pod \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.689139 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpxv7\" (UniqueName: \"kubernetes.io/projected/c41402cd-e31f-47ed-ba7a-1f352c1c819c-kube-api-access-bpxv7\") pod \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\" (UID: \"c41402cd-e31f-47ed-ba7a-1f352c1c819c\") " Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.689881 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c41402cd-e31f-47ed-ba7a-1f352c1c819c-logs" (OuterVolumeSpecName: "logs") pod "c41402cd-e31f-47ed-ba7a-1f352c1c819c" (UID: "c41402cd-e31f-47ed-ba7a-1f352c1c819c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.690545 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c41402cd-e31f-47ed-ba7a-1f352c1c819c-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.694314 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c41402cd-e31f-47ed-ba7a-1f352c1c819c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c41402cd-e31f-47ed-ba7a-1f352c1c819c" (UID: "c41402cd-e31f-47ed-ba7a-1f352c1c819c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.696586 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c41402cd-e31f-47ed-ba7a-1f352c1c819c-kube-api-access-bpxv7" (OuterVolumeSpecName: "kube-api-access-bpxv7") pod "c41402cd-e31f-47ed-ba7a-1f352c1c819c" (UID: "c41402cd-e31f-47ed-ba7a-1f352c1c819c"). InnerVolumeSpecName "kube-api-access-bpxv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.715861 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-config-data" (OuterVolumeSpecName: "config-data") pod "c41402cd-e31f-47ed-ba7a-1f352c1c819c" (UID: "c41402cd-e31f-47ed-ba7a-1f352c1c819c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.715873 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-scripts" (OuterVolumeSpecName: "scripts") pod "c41402cd-e31f-47ed-ba7a-1f352c1c819c" (UID: "c41402cd-e31f-47ed-ba7a-1f352c1c819c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.792671 4945 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c41402cd-e31f-47ed-ba7a-1f352c1c819c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.792726 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.792744 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpxv7\" (UniqueName: \"kubernetes.io/projected/c41402cd-e31f-47ed-ba7a-1f352c1c819c-kube-api-access-bpxv7\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.792765 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c41402cd-e31f-47ed-ba7a-1f352c1c819c-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.945029 4945 scope.go:117] "RemoveContainer" containerID="91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.965255 4945 scope.go:117] "RemoveContainer" containerID="828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d" Dec 06 08:28:05 crc kubenswrapper[4945]: E1206 08:28:05.965774 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d\": container with ID starting with 828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d not found: ID does not exist" containerID="828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.965810 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d"} err="failed to get container status \"828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d\": rpc error: code = NotFound desc = could not find container \"828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d\": container with ID starting with 828a422869ed2498f91960544108d31a2bde578a0f25d01f602c61438d86666d not found: ID does not exist" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.965835 4945 scope.go:117] "RemoveContainer" containerID="91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c" Dec 06 08:28:05 crc kubenswrapper[4945]: E1206 08:28:05.966216 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c\": container with ID starting with 91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c not found: ID does not exist" containerID="91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c" Dec 06 08:28:05 crc kubenswrapper[4945]: I1206 08:28:05.966323 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c"} err="failed to get container status \"91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c\": rpc error: code = NotFound desc = could not find container \"91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c\": container with ID starting with 91550fb372a86a9bad84ef6e65d36d4bb32fb8530c2eeb60fc6a3c43ab0b3f7c not found: ID does not exist" Dec 06 08:28:06 crc kubenswrapper[4945]: I1206 08:28:06.030814 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-748744fc9c-m2ggc"] Dec 06 08:28:06 crc kubenswrapper[4945]: I1206 08:28:06.042980 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-748744fc9c-m2ggc"] Dec 06 08:28:06 crc kubenswrapper[4945]: I1206 08:28:06.970266 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" path="/var/lib/kubelet/pods/c41402cd-e31f-47ed-ba7a-1f352c1c819c/volumes" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.490592 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64775b6477-kp8ql"] Dec 06 08:28:18 crc kubenswrapper[4945]: E1206 08:28:18.491520 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon-log" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.491538 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon-log" Dec 06 08:28:18 crc kubenswrapper[4945]: E1206 08:28:18.491555 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.491560 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon" Dec 06 08:28:18 crc kubenswrapper[4945]: E1206 08:28:18.491583 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e872eee-469e-480b-9eca-c89aa6a1be40" containerName="horizon-log" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.491590 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e872eee-469e-480b-9eca-c89aa6a1be40" containerName="horizon-log" Dec 06 08:28:18 crc kubenswrapper[4945]: E1206 08:28:18.491614 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e872eee-469e-480b-9eca-c89aa6a1be40" containerName="horizon" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.491622 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e872eee-469e-480b-9eca-c89aa6a1be40" containerName="horizon" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.491813 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.491825 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e872eee-469e-480b-9eca-c89aa6a1be40" containerName="horizon-log" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.491842 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e872eee-469e-480b-9eca-c89aa6a1be40" containerName="horizon" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.491855 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c41402cd-e31f-47ed-ba7a-1f352c1c819c" containerName="horizon-log" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.493008 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.557126 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-scripts\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.557559 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-config-data\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.557723 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2fwk\" (UniqueName: \"kubernetes.io/projected/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-kube-api-access-v2fwk\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.557816 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-horizon-secret-key\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.557861 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-logs\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.560846 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64775b6477-kp8ql"] Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.660005 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2fwk\" (UniqueName: \"kubernetes.io/projected/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-kube-api-access-v2fwk\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.660056 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-horizon-secret-key\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.660079 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-logs\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.660186 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-scripts\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.660236 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-config-data\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.660726 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-logs\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.661035 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-scripts\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.662314 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-config-data\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.673109 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-horizon-secret-key\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.676919 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2fwk\" (UniqueName: \"kubernetes.io/projected/e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9-kube-api-access-v2fwk\") pod \"horizon-64775b6477-kp8ql\" (UID: \"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9\") " pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:18 crc kubenswrapper[4945]: I1206 08:28:18.821734 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:19 crc kubenswrapper[4945]: I1206 08:28:19.363082 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64775b6477-kp8ql"] Dec 06 08:28:19 crc kubenswrapper[4945]: I1206 08:28:19.838886 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64775b6477-kp8ql" event={"ID":"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9","Type":"ContainerStarted","Data":"b1974b6236c1420b9de2fbedbfaa325a433501bddd56939f270ac45b2942b541"} Dec 06 08:28:19 crc kubenswrapper[4945]: I1206 08:28:19.839382 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64775b6477-kp8ql" event={"ID":"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9","Type":"ContainerStarted","Data":"e176469d3eca51d41125c7d0782fb8236a9672fa00f554ca04db87a92dc74f11"} Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.112530 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-nzd5c"] Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.114038 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-nzd5c" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.128294 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-nzd5c"] Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.188934 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsj22\" (UniqueName: \"kubernetes.io/projected/21cfae16-fc47-4e70-b4a7-42d57390890d-kube-api-access-qsj22\") pod \"heat-db-create-nzd5c\" (UID: \"21cfae16-fc47-4e70-b4a7-42d57390890d\") " pod="openstack/heat-db-create-nzd5c" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.189056 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21cfae16-fc47-4e70-b4a7-42d57390890d-operator-scripts\") pod \"heat-db-create-nzd5c\" (UID: \"21cfae16-fc47-4e70-b4a7-42d57390890d\") " pod="openstack/heat-db-create-nzd5c" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.215607 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-7256-account-create-update-9jbzh"] Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.216854 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-7256-account-create-update-9jbzh" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.219238 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.227202 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-7256-account-create-update-9jbzh"] Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.290888 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21cfae16-fc47-4e70-b4a7-42d57390890d-operator-scripts\") pod \"heat-db-create-nzd5c\" (UID: \"21cfae16-fc47-4e70-b4a7-42d57390890d\") " pod="openstack/heat-db-create-nzd5c" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.291333 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsj22\" (UniqueName: \"kubernetes.io/projected/21cfae16-fc47-4e70-b4a7-42d57390890d-kube-api-access-qsj22\") pod \"heat-db-create-nzd5c\" (UID: \"21cfae16-fc47-4e70-b4a7-42d57390890d\") " pod="openstack/heat-db-create-nzd5c" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.291628 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjxrx\" (UniqueName: \"kubernetes.io/projected/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-kube-api-access-cjxrx\") pod \"heat-7256-account-create-update-9jbzh\" (UID: \"b7d5cf50-be90-4e38-83f8-a42bddf21ec8\") " pod="openstack/heat-7256-account-create-update-9jbzh" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.291815 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-operator-scripts\") pod \"heat-7256-account-create-update-9jbzh\" (UID: \"b7d5cf50-be90-4e38-83f8-a42bddf21ec8\") " pod="openstack/heat-7256-account-create-update-9jbzh" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.291984 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21cfae16-fc47-4e70-b4a7-42d57390890d-operator-scripts\") pod \"heat-db-create-nzd5c\" (UID: \"21cfae16-fc47-4e70-b4a7-42d57390890d\") " pod="openstack/heat-db-create-nzd5c" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.317046 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsj22\" (UniqueName: \"kubernetes.io/projected/21cfae16-fc47-4e70-b4a7-42d57390890d-kube-api-access-qsj22\") pod \"heat-db-create-nzd5c\" (UID: \"21cfae16-fc47-4e70-b4a7-42d57390890d\") " pod="openstack/heat-db-create-nzd5c" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.393217 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjxrx\" (UniqueName: \"kubernetes.io/projected/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-kube-api-access-cjxrx\") pod \"heat-7256-account-create-update-9jbzh\" (UID: \"b7d5cf50-be90-4e38-83f8-a42bddf21ec8\") " pod="openstack/heat-7256-account-create-update-9jbzh" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.393292 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-operator-scripts\") pod \"heat-7256-account-create-update-9jbzh\" (UID: \"b7d5cf50-be90-4e38-83f8-a42bddf21ec8\") " pod="openstack/heat-7256-account-create-update-9jbzh" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.394020 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-operator-scripts\") pod \"heat-7256-account-create-update-9jbzh\" (UID: \"b7d5cf50-be90-4e38-83f8-a42bddf21ec8\") " pod="openstack/heat-7256-account-create-update-9jbzh" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.408885 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjxrx\" (UniqueName: \"kubernetes.io/projected/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-kube-api-access-cjxrx\") pod \"heat-7256-account-create-update-9jbzh\" (UID: \"b7d5cf50-be90-4e38-83f8-a42bddf21ec8\") " pod="openstack/heat-7256-account-create-update-9jbzh" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.447802 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-nzd5c" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.541040 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-7256-account-create-update-9jbzh" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.851387 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64775b6477-kp8ql" event={"ID":"e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9","Type":"ContainerStarted","Data":"620aa0be1c1f6d775bd003446b74aa71b5e26cc3465786b3965de64b0c884706"} Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.888304 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-64775b6477-kp8ql" podStartSLOduration=2.888252139 podStartE2EDuration="2.888252139s" podCreationTimestamp="2025-12-06 08:28:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:28:20.866464976 +0000 UTC m=+5714.321326010" watchObservedRunningTime="2025-12-06 08:28:20.888252139 +0000 UTC m=+5714.343113183" Dec 06 08:28:20 crc kubenswrapper[4945]: I1206 08:28:20.936179 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-nzd5c"] Dec 06 08:28:21 crc kubenswrapper[4945]: I1206 08:28:21.023731 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-7256-account-create-update-9jbzh"] Dec 06 08:28:21 crc kubenswrapper[4945]: W1206 08:28:21.030053 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5cf50_be90_4e38_83f8_a42bddf21ec8.slice/crio-a7880d4ff7af885b97eb633c4f7caae8d83f35648c5a3612c497ae29321a5b5a WatchSource:0}: Error finding container a7880d4ff7af885b97eb633c4f7caae8d83f35648c5a3612c497ae29321a5b5a: Status 404 returned error can't find the container with id a7880d4ff7af885b97eb633c4f7caae8d83f35648c5a3612c497ae29321a5b5a Dec 06 08:28:21 crc kubenswrapper[4945]: I1206 08:28:21.861977 4945 generic.go:334] "Generic (PLEG): container finished" podID="b7d5cf50-be90-4e38-83f8-a42bddf21ec8" containerID="1a55017d1f289cc210cd6b209f79e3b08ac403dfa99c19d1d5f30572db9c2bc0" exitCode=0 Dec 06 08:28:21 crc kubenswrapper[4945]: I1206 08:28:21.862224 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-7256-account-create-update-9jbzh" event={"ID":"b7d5cf50-be90-4e38-83f8-a42bddf21ec8","Type":"ContainerDied","Data":"1a55017d1f289cc210cd6b209f79e3b08ac403dfa99c19d1d5f30572db9c2bc0"} Dec 06 08:28:21 crc kubenswrapper[4945]: I1206 08:28:21.862358 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-7256-account-create-update-9jbzh" event={"ID":"b7d5cf50-be90-4e38-83f8-a42bddf21ec8","Type":"ContainerStarted","Data":"a7880d4ff7af885b97eb633c4f7caae8d83f35648c5a3612c497ae29321a5b5a"} Dec 06 08:28:21 crc kubenswrapper[4945]: I1206 08:28:21.870554 4945 generic.go:334] "Generic (PLEG): container finished" podID="21cfae16-fc47-4e70-b4a7-42d57390890d" containerID="d997631c28d3a68669756655cf73d65383aa2afee24a7dd42356460799b22d44" exitCode=0 Dec 06 08:28:21 crc kubenswrapper[4945]: I1206 08:28:21.870637 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-nzd5c" event={"ID":"21cfae16-fc47-4e70-b4a7-42d57390890d","Type":"ContainerDied","Data":"d997631c28d3a68669756655cf73d65383aa2afee24a7dd42356460799b22d44"} Dec 06 08:28:21 crc kubenswrapper[4945]: I1206 08:28:21.870809 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-nzd5c" event={"ID":"21cfae16-fc47-4e70-b4a7-42d57390890d","Type":"ContainerStarted","Data":"0cb447e0ca8c8196bde86c0ac723c935c64fa38513f439007b3df636c26c9d72"} Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.292928 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-7256-account-create-update-9jbzh" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.305495 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-nzd5c" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.451930 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-operator-scripts\") pod \"b7d5cf50-be90-4e38-83f8-a42bddf21ec8\" (UID: \"b7d5cf50-be90-4e38-83f8-a42bddf21ec8\") " Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.452008 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjxrx\" (UniqueName: \"kubernetes.io/projected/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-kube-api-access-cjxrx\") pod \"b7d5cf50-be90-4e38-83f8-a42bddf21ec8\" (UID: \"b7d5cf50-be90-4e38-83f8-a42bddf21ec8\") " Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.452139 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsj22\" (UniqueName: \"kubernetes.io/projected/21cfae16-fc47-4e70-b4a7-42d57390890d-kube-api-access-qsj22\") pod \"21cfae16-fc47-4e70-b4a7-42d57390890d\" (UID: \"21cfae16-fc47-4e70-b4a7-42d57390890d\") " Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.452264 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21cfae16-fc47-4e70-b4a7-42d57390890d-operator-scripts\") pod \"21cfae16-fc47-4e70-b4a7-42d57390890d\" (UID: \"21cfae16-fc47-4e70-b4a7-42d57390890d\") " Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.452978 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21cfae16-fc47-4e70-b4a7-42d57390890d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "21cfae16-fc47-4e70-b4a7-42d57390890d" (UID: "21cfae16-fc47-4e70-b4a7-42d57390890d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.453371 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b7d5cf50-be90-4e38-83f8-a42bddf21ec8" (UID: "b7d5cf50-be90-4e38-83f8-a42bddf21ec8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.457761 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-kube-api-access-cjxrx" (OuterVolumeSpecName: "kube-api-access-cjxrx") pod "b7d5cf50-be90-4e38-83f8-a42bddf21ec8" (UID: "b7d5cf50-be90-4e38-83f8-a42bddf21ec8"). InnerVolumeSpecName "kube-api-access-cjxrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.468386 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21cfae16-fc47-4e70-b4a7-42d57390890d-kube-api-access-qsj22" (OuterVolumeSpecName: "kube-api-access-qsj22") pod "21cfae16-fc47-4e70-b4a7-42d57390890d" (UID: "21cfae16-fc47-4e70-b4a7-42d57390890d"). InnerVolumeSpecName "kube-api-access-qsj22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.554540 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsj22\" (UniqueName: \"kubernetes.io/projected/21cfae16-fc47-4e70-b4a7-42d57390890d-kube-api-access-qsj22\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.554571 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21cfae16-fc47-4e70-b4a7-42d57390890d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.554586 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.554615 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjxrx\" (UniqueName: \"kubernetes.io/projected/b7d5cf50-be90-4e38-83f8-a42bddf21ec8-kube-api-access-cjxrx\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.894398 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-7256-account-create-update-9jbzh" event={"ID":"b7d5cf50-be90-4e38-83f8-a42bddf21ec8","Type":"ContainerDied","Data":"a7880d4ff7af885b97eb633c4f7caae8d83f35648c5a3612c497ae29321a5b5a"} Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.894669 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7880d4ff7af885b97eb633c4f7caae8d83f35648c5a3612c497ae29321a5b5a" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.894720 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-7256-account-create-update-9jbzh" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.905158 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-nzd5c" event={"ID":"21cfae16-fc47-4e70-b4a7-42d57390890d","Type":"ContainerDied","Data":"0cb447e0ca8c8196bde86c0ac723c935c64fa38513f439007b3df636c26c9d72"} Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.905199 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cb447e0ca8c8196bde86c0ac723c935c64fa38513f439007b3df636c26c9d72" Dec 06 08:28:23 crc kubenswrapper[4945]: I1206 08:28:23.905265 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-nzd5c" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.798103 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-sb9hp"] Dec 06 08:28:25 crc kubenswrapper[4945]: E1206 08:28:25.798760 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d5cf50-be90-4e38-83f8-a42bddf21ec8" containerName="mariadb-account-create-update" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.798786 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d5cf50-be90-4e38-83f8-a42bddf21ec8" containerName="mariadb-account-create-update" Dec 06 08:28:25 crc kubenswrapper[4945]: E1206 08:28:25.798862 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21cfae16-fc47-4e70-b4a7-42d57390890d" containerName="mariadb-database-create" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.798874 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="21cfae16-fc47-4e70-b4a7-42d57390890d" containerName="mariadb-database-create" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.799221 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7d5cf50-be90-4e38-83f8-a42bddf21ec8" containerName="mariadb-account-create-update" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.799250 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="21cfae16-fc47-4e70-b4a7-42d57390890d" containerName="mariadb-database-create" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.800413 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.827238 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.828568 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-cb6cf" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.870535 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-sb9hp"] Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.902499 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-config-data\") pod \"heat-db-sync-sb9hp\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.902576 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-combined-ca-bundle\") pod \"heat-db-sync-sb9hp\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:25 crc kubenswrapper[4945]: I1206 08:28:25.902957 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp6tn\" (UniqueName: \"kubernetes.io/projected/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-kube-api-access-wp6tn\") pod \"heat-db-sync-sb9hp\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:26 crc kubenswrapper[4945]: I1206 08:28:26.005190 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-combined-ca-bundle\") pod \"heat-db-sync-sb9hp\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:26 crc kubenswrapper[4945]: I1206 08:28:26.005361 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp6tn\" (UniqueName: \"kubernetes.io/projected/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-kube-api-access-wp6tn\") pod \"heat-db-sync-sb9hp\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:26 crc kubenswrapper[4945]: I1206 08:28:26.005499 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-config-data\") pod \"heat-db-sync-sb9hp\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:26 crc kubenswrapper[4945]: I1206 08:28:26.011327 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-combined-ca-bundle\") pod \"heat-db-sync-sb9hp\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:26 crc kubenswrapper[4945]: I1206 08:28:26.013143 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-config-data\") pod \"heat-db-sync-sb9hp\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:26 crc kubenswrapper[4945]: I1206 08:28:26.022637 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp6tn\" (UniqueName: \"kubernetes.io/projected/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-kube-api-access-wp6tn\") pod \"heat-db-sync-sb9hp\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:26 crc kubenswrapper[4945]: I1206 08:28:26.161381 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:26 crc kubenswrapper[4945]: I1206 08:28:26.641891 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-sb9hp"] Dec 06 08:28:26 crc kubenswrapper[4945]: I1206 08:28:26.944935 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-sb9hp" event={"ID":"5a1010c7-ceeb-47e1-b51d-a5d20bea055b","Type":"ContainerStarted","Data":"003c3b9e3b6fc49bd194f588646f0eba71aa763a11a12d10827901a80b82929a"} Dec 06 08:28:28 crc kubenswrapper[4945]: I1206 08:28:28.822688 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:28 crc kubenswrapper[4945]: I1206 08:28:28.823557 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:35 crc kubenswrapper[4945]: I1206 08:28:35.017067 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-sb9hp" event={"ID":"5a1010c7-ceeb-47e1-b51d-a5d20bea055b","Type":"ContainerStarted","Data":"aa45bbe5ab5ab0e551d313b87afadbb4ad810c3ce47286c9d3904487dd18a4b6"} Dec 06 08:28:35 crc kubenswrapper[4945]: I1206 08:28:35.040810 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-sb9hp" podStartSLOduration=2.628943144 podStartE2EDuration="10.040783956s" podCreationTimestamp="2025-12-06 08:28:25 +0000 UTC" firstStartedPulling="2025-12-06 08:28:26.649238142 +0000 UTC m=+5720.104099186" lastFinishedPulling="2025-12-06 08:28:34.061078954 +0000 UTC m=+5727.515939998" observedRunningTime="2025-12-06 08:28:35.033563893 +0000 UTC m=+5728.488424957" watchObservedRunningTime="2025-12-06 08:28:35.040783956 +0000 UTC m=+5728.495645020" Dec 06 08:28:36 crc kubenswrapper[4945]: I1206 08:28:36.031523 4945 generic.go:334] "Generic (PLEG): container finished" podID="5a1010c7-ceeb-47e1-b51d-a5d20bea055b" containerID="aa45bbe5ab5ab0e551d313b87afadbb4ad810c3ce47286c9d3904487dd18a4b6" exitCode=0 Dec 06 08:28:36 crc kubenswrapper[4945]: I1206 08:28:36.031867 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-sb9hp" event={"ID":"5a1010c7-ceeb-47e1-b51d-a5d20bea055b","Type":"ContainerDied","Data":"aa45bbe5ab5ab0e551d313b87afadbb4ad810c3ce47286c9d3904487dd18a4b6"} Dec 06 08:28:36 crc kubenswrapper[4945]: I1206 08:28:36.050990 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-v879n"] Dec 06 08:28:36 crc kubenswrapper[4945]: I1206 08:28:36.060027 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-39b5-account-create-update-6l9r9"] Dec 06 08:28:36 crc kubenswrapper[4945]: I1206 08:28:36.068616 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-39b5-account-create-update-6l9r9"] Dec 06 08:28:36 crc kubenswrapper[4945]: I1206 08:28:36.077120 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-v879n"] Dec 06 08:28:36 crc kubenswrapper[4945]: I1206 08:28:36.966907 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9" path="/var/lib/kubelet/pods/9c2ce74f-0f26-4db5-b6b4-10ebb3ad80f9/volumes" Dec 06 08:28:36 crc kubenswrapper[4945]: I1206 08:28:36.967499 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de99182f-85eb-43ec-a5a5-c5ec7779b399" path="/var/lib/kubelet/pods/de99182f-85eb-43ec-a5a5-c5ec7779b399/volumes" Dec 06 08:28:37 crc kubenswrapper[4945]: I1206 08:28:37.398435 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:37 crc kubenswrapper[4945]: I1206 08:28:37.435728 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-combined-ca-bundle\") pod \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " Dec 06 08:28:37 crc kubenswrapper[4945]: I1206 08:28:37.435882 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp6tn\" (UniqueName: \"kubernetes.io/projected/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-kube-api-access-wp6tn\") pod \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " Dec 06 08:28:37 crc kubenswrapper[4945]: I1206 08:28:37.435960 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-config-data\") pod \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\" (UID: \"5a1010c7-ceeb-47e1-b51d-a5d20bea055b\") " Dec 06 08:28:37 crc kubenswrapper[4945]: I1206 08:28:37.441666 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-kube-api-access-wp6tn" (OuterVolumeSpecName: "kube-api-access-wp6tn") pod "5a1010c7-ceeb-47e1-b51d-a5d20bea055b" (UID: "5a1010c7-ceeb-47e1-b51d-a5d20bea055b"). InnerVolumeSpecName "kube-api-access-wp6tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:28:37 crc kubenswrapper[4945]: I1206 08:28:37.460862 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a1010c7-ceeb-47e1-b51d-a5d20bea055b" (UID: "5a1010c7-ceeb-47e1-b51d-a5d20bea055b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:28:37 crc kubenswrapper[4945]: I1206 08:28:37.504998 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-config-data" (OuterVolumeSpecName: "config-data") pod "5a1010c7-ceeb-47e1-b51d-a5d20bea055b" (UID: "5a1010c7-ceeb-47e1-b51d-a5d20bea055b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:28:37 crc kubenswrapper[4945]: I1206 08:28:37.538414 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp6tn\" (UniqueName: \"kubernetes.io/projected/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-kube-api-access-wp6tn\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:37 crc kubenswrapper[4945]: I1206 08:28:37.538482 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:37 crc kubenswrapper[4945]: I1206 08:28:37.538493 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1010c7-ceeb-47e1-b51d-a5d20bea055b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:28:38 crc kubenswrapper[4945]: I1206 08:28:38.053930 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-sb9hp" event={"ID":"5a1010c7-ceeb-47e1-b51d-a5d20bea055b","Type":"ContainerDied","Data":"003c3b9e3b6fc49bd194f588646f0eba71aa763a11a12d10827901a80b82929a"} Dec 06 08:28:38 crc kubenswrapper[4945]: I1206 08:28:38.054260 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="003c3b9e3b6fc49bd194f588646f0eba71aa763a11a12d10827901a80b82929a" Dec 06 08:28:38 crc kubenswrapper[4945]: I1206 08:28:38.054348 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-sb9hp" Dec 06 08:28:38 crc kubenswrapper[4945]: I1206 08:28:38.796563 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:28:38 crc kubenswrapper[4945]: I1206 08:28:38.796627 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.065392 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5445fc8644-scpcx"] Dec 06 08:28:39 crc kubenswrapper[4945]: E1206 08:28:39.066078 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1010c7-ceeb-47e1-b51d-a5d20bea055b" containerName="heat-db-sync" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.066163 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1010c7-ceeb-47e1-b51d-a5d20bea055b" containerName="heat-db-sync" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.066422 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1010c7-ceeb-47e1-b51d-a5d20bea055b" containerName="heat-db-sync" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.067114 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.075759 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.076049 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-cb6cf" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.076336 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.113977 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5445fc8644-scpcx"] Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.169438 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdxln\" (UniqueName: \"kubernetes.io/projected/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-kube-api-access-wdxln\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.169595 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-combined-ca-bundle\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.169625 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-config-data\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.169694 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-config-data-custom\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.225300 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-779bcf995f-tvh72"] Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.229039 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.231597 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.240840 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-779bcf995f-tvh72"] Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.257642 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-694fddd77-dwqw7"] Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.258947 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.261773 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.272567 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b92b2c6-fac7-4da7-9291-ffcac487b85f-config-data\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.272636 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b92b2c6-fac7-4da7-9291-ffcac487b85f-config-data-custom\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.272678 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdxln\" (UniqueName: \"kubernetes.io/projected/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-kube-api-access-wdxln\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.272747 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b92b2c6-fac7-4da7-9291-ffcac487b85f-combined-ca-bundle\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.272775 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6shxw\" (UniqueName: \"kubernetes.io/projected/0b92b2c6-fac7-4da7-9291-ffcac487b85f-kube-api-access-6shxw\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.272813 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-combined-ca-bundle\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.272830 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-config-data\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.272879 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-config-data-custom\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.273851 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-694fddd77-dwqw7"] Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.280116 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-combined-ca-bundle\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.289622 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-config-data\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.290415 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-config-data-custom\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.308035 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdxln\" (UniqueName: \"kubernetes.io/projected/97d626d5-7adf-40dd-9b6a-cf361f3c34cf-kube-api-access-wdxln\") pod \"heat-engine-5445fc8644-scpcx\" (UID: \"97d626d5-7adf-40dd-9b6a-cf361f3c34cf\") " pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.375881 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms2mz\" (UniqueName: \"kubernetes.io/projected/7cd86064-9374-494e-90b6-dac599098554-kube-api-access-ms2mz\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.375943 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cd86064-9374-494e-90b6-dac599098554-config-data-custom\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.376000 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b92b2c6-fac7-4da7-9291-ffcac487b85f-config-data\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.376029 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b92b2c6-fac7-4da7-9291-ffcac487b85f-config-data-custom\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.376328 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b92b2c6-fac7-4da7-9291-ffcac487b85f-combined-ca-bundle\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.376446 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6shxw\" (UniqueName: \"kubernetes.io/projected/0b92b2c6-fac7-4da7-9291-ffcac487b85f-kube-api-access-6shxw\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.376495 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd86064-9374-494e-90b6-dac599098554-combined-ca-bundle\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.376525 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cd86064-9374-494e-90b6-dac599098554-config-data\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.380990 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b92b2c6-fac7-4da7-9291-ffcac487b85f-combined-ca-bundle\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.381099 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b92b2c6-fac7-4da7-9291-ffcac487b85f-config-data\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.381949 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b92b2c6-fac7-4da7-9291-ffcac487b85f-config-data-custom\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.389808 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.396071 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6shxw\" (UniqueName: \"kubernetes.io/projected/0b92b2c6-fac7-4da7-9291-ffcac487b85f-kube-api-access-6shxw\") pod \"heat-cfnapi-779bcf995f-tvh72\" (UID: \"0b92b2c6-fac7-4da7-9291-ffcac487b85f\") " pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.478239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd86064-9374-494e-90b6-dac599098554-combined-ca-bundle\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.478302 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cd86064-9374-494e-90b6-dac599098554-config-data\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.478341 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms2mz\" (UniqueName: \"kubernetes.io/projected/7cd86064-9374-494e-90b6-dac599098554-kube-api-access-ms2mz\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.478389 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cd86064-9374-494e-90b6-dac599098554-config-data-custom\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.486570 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cd86064-9374-494e-90b6-dac599098554-config-data\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.488490 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd86064-9374-494e-90b6-dac599098554-combined-ca-bundle\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.489293 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cd86064-9374-494e-90b6-dac599098554-config-data-custom\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.502922 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms2mz\" (UniqueName: \"kubernetes.io/projected/7cd86064-9374-494e-90b6-dac599098554-kube-api-access-ms2mz\") pod \"heat-api-694fddd77-dwqw7\" (UID: \"7cd86064-9374-494e-90b6-dac599098554\") " pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.557805 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:39 crc kubenswrapper[4945]: I1206 08:28:39.672847 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:40 crc kubenswrapper[4945]: I1206 08:28:40.084354 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5445fc8644-scpcx"] Dec 06 08:28:40 crc kubenswrapper[4945]: I1206 08:28:40.299418 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-779bcf995f-tvh72"] Dec 06 08:28:40 crc kubenswrapper[4945]: I1206 08:28:40.355026 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-694fddd77-dwqw7"] Dec 06 08:28:40 crc kubenswrapper[4945]: W1206 08:28:40.355548 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cd86064_9374_494e_90b6_dac599098554.slice/crio-3d1e869b9f0012dbc7f2ab3d8a098b8f2541a2bf12336f89e45cffebc7f75a3f WatchSource:0}: Error finding container 3d1e869b9f0012dbc7f2ab3d8a098b8f2541a2bf12336f89e45cffebc7f75a3f: Status 404 returned error can't find the container with id 3d1e869b9f0012dbc7f2ab3d8a098b8f2541a2bf12336f89e45cffebc7f75a3f Dec 06 08:28:41 crc kubenswrapper[4945]: I1206 08:28:41.090016 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-694fddd77-dwqw7" event={"ID":"7cd86064-9374-494e-90b6-dac599098554","Type":"ContainerStarted","Data":"3d1e869b9f0012dbc7f2ab3d8a098b8f2541a2bf12336f89e45cffebc7f75a3f"} Dec 06 08:28:41 crc kubenswrapper[4945]: I1206 08:28:41.092404 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5445fc8644-scpcx" event={"ID":"97d626d5-7adf-40dd-9b6a-cf361f3c34cf","Type":"ContainerStarted","Data":"3d0d5296c0bb6a988a917119f3594ec67a847f2c6ef4feea401b61a8cb513ecb"} Dec 06 08:28:41 crc kubenswrapper[4945]: I1206 08:28:41.092459 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5445fc8644-scpcx" event={"ID":"97d626d5-7adf-40dd-9b6a-cf361f3c34cf","Type":"ContainerStarted","Data":"41145d24c6825b84d38568caf025508aa18a44eb4a492b6498446e3d5c8d40e6"} Dec 06 08:28:41 crc kubenswrapper[4945]: I1206 08:28:41.093694 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:28:41 crc kubenswrapper[4945]: I1206 08:28:41.098155 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-779bcf995f-tvh72" event={"ID":"0b92b2c6-fac7-4da7-9291-ffcac487b85f","Type":"ContainerStarted","Data":"037902f0a9667ffe5421903569a1260d20b9b206c7ff0d89fc47c6c7eab331a3"} Dec 06 08:28:41 crc kubenswrapper[4945]: I1206 08:28:41.105133 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:41 crc kubenswrapper[4945]: I1206 08:28:41.109225 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5445fc8644-scpcx" podStartSLOduration=2.109202305 podStartE2EDuration="2.109202305s" podCreationTimestamp="2025-12-06 08:28:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:28:41.108825295 +0000 UTC m=+5734.563686349" watchObservedRunningTime="2025-12-06 08:28:41.109202305 +0000 UTC m=+5734.564063349" Dec 06 08:28:43 crc kubenswrapper[4945]: I1206 08:28:43.025914 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-64775b6477-kp8ql" Dec 06 08:28:43 crc kubenswrapper[4945]: I1206 08:28:43.102045 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67bcf67b9f-fln2w"] Dec 06 08:28:43 crc kubenswrapper[4945]: I1206 08:28:43.102496 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67bcf67b9f-fln2w" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon-log" containerID="cri-o://e490063ebf7044297abdd8d76cff3a7686ad4ed811ca3a265e5ef91372cfc76b" gracePeriod=30 Dec 06 08:28:43 crc kubenswrapper[4945]: I1206 08:28:43.102827 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67bcf67b9f-fln2w" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon" containerID="cri-o://73f4abd92b9c37ed61a7ad565c7e25a5ca8dd2322dc464962eafe564a1fd020e" gracePeriod=30 Dec 06 08:28:43 crc kubenswrapper[4945]: I1206 08:28:43.134971 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-694fddd77-dwqw7" event={"ID":"7cd86064-9374-494e-90b6-dac599098554","Type":"ContainerStarted","Data":"69e2f7f5533c3d7aa02b3465a50a43f60273f9eb9ecb17d701bec00065adcc73"} Dec 06 08:28:43 crc kubenswrapper[4945]: I1206 08:28:43.135452 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:43 crc kubenswrapper[4945]: I1206 08:28:43.139544 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-779bcf995f-tvh72" event={"ID":"0b92b2c6-fac7-4da7-9291-ffcac487b85f","Type":"ContainerStarted","Data":"4748b2c65d138dbcae2ff8cae3b23b07248fd36f5a7f9e4277a00e24ec3c2393"} Dec 06 08:28:43 crc kubenswrapper[4945]: I1206 08:28:43.139711 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:43 crc kubenswrapper[4945]: I1206 08:28:43.183266 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-694fddd77-dwqw7" podStartSLOduration=2.563763498 podStartE2EDuration="4.183246806s" podCreationTimestamp="2025-12-06 08:28:39 +0000 UTC" firstStartedPulling="2025-12-06 08:28:40.358387358 +0000 UTC m=+5733.813248402" lastFinishedPulling="2025-12-06 08:28:41.977870666 +0000 UTC m=+5735.432731710" observedRunningTime="2025-12-06 08:28:43.166068126 +0000 UTC m=+5736.620929170" watchObservedRunningTime="2025-12-06 08:28:43.183246806 +0000 UTC m=+5736.638107850" Dec 06 08:28:43 crc kubenswrapper[4945]: I1206 08:28:43.188492 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-779bcf995f-tvh72" podStartSLOduration=2.516674918 podStartE2EDuration="4.188478526s" podCreationTimestamp="2025-12-06 08:28:39 +0000 UTC" firstStartedPulling="2025-12-06 08:28:40.302568874 +0000 UTC m=+5733.757429918" lastFinishedPulling="2025-12-06 08:28:41.974372492 +0000 UTC m=+5735.429233526" observedRunningTime="2025-12-06 08:28:43.185108786 +0000 UTC m=+5736.639969830" watchObservedRunningTime="2025-12-06 08:28:43.188478526 +0000 UTC m=+5736.643339570" Dec 06 08:28:47 crc kubenswrapper[4945]: I1206 08:28:47.043252 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-t2htt"] Dec 06 08:28:47 crc kubenswrapper[4945]: I1206 08:28:47.056423 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-t2htt"] Dec 06 08:28:47 crc kubenswrapper[4945]: I1206 08:28:47.176902 4945 generic.go:334] "Generic (PLEG): container finished" podID="7547017f-a981-4309-979f-6776230b2198" containerID="73f4abd92b9c37ed61a7ad565c7e25a5ca8dd2322dc464962eafe564a1fd020e" exitCode=0 Dec 06 08:28:47 crc kubenswrapper[4945]: I1206 08:28:47.177064 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bcf67b9f-fln2w" event={"ID":"7547017f-a981-4309-979f-6776230b2198","Type":"ContainerDied","Data":"73f4abd92b9c37ed61a7ad565c7e25a5ca8dd2322dc464962eafe564a1fd020e"} Dec 06 08:28:48 crc kubenswrapper[4945]: I1206 08:28:48.967019 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee0e0648-81a5-4118-9e0f-53ad852f145e" path="/var/lib/kubelet/pods/ee0e0648-81a5-4118-9e0f-53ad852f145e/volumes" Dec 06 08:28:50 crc kubenswrapper[4945]: I1206 08:28:50.970923 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-694fddd77-dwqw7" Dec 06 08:28:51 crc kubenswrapper[4945]: I1206 08:28:51.063154 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-779bcf995f-tvh72" Dec 06 08:28:51 crc kubenswrapper[4945]: I1206 08:28:51.064568 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67bcf67b9f-fln2w" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.92:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.92:8080: connect: connection refused" Dec 06 08:28:59 crc kubenswrapper[4945]: I1206 08:28:59.417259 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5445fc8644-scpcx" Dec 06 08:29:01 crc kubenswrapper[4945]: I1206 08:29:01.064639 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67bcf67b9f-fln2w" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.92:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.92:8080: connect: connection refused" Dec 06 08:29:08 crc kubenswrapper[4945]: I1206 08:29:08.795573 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:29:08 crc kubenswrapper[4945]: I1206 08:29:08.796188 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.288482 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl"] Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.318955 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl"] Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.319139 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.324271 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.400244 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.400363 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl8l7\" (UniqueName: \"kubernetes.io/projected/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-kube-api-access-jl8l7\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.400402 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.501591 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.501687 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl8l7\" (UniqueName: \"kubernetes.io/projected/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-kube-api-access-jl8l7\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.501726 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.502259 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.502434 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.520359 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl8l7\" (UniqueName: \"kubernetes.io/projected/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-kube-api-access-jl8l7\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:09 crc kubenswrapper[4945]: I1206 08:29:09.647893 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:10 crc kubenswrapper[4945]: W1206 08:29:10.152507 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6cd545a_16b7_49ca_bcb9_3a733c7f4499.slice/crio-b04355382e4c51f8c1ce2fdb9abbdfe28af4abb8e93ebbb92465e073840d64af WatchSource:0}: Error finding container b04355382e4c51f8c1ce2fdb9abbdfe28af4abb8e93ebbb92465e073840d64af: Status 404 returned error can't find the container with id b04355382e4c51f8c1ce2fdb9abbdfe28af4abb8e93ebbb92465e073840d64af Dec 06 08:29:10 crc kubenswrapper[4945]: I1206 08:29:10.154926 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl"] Dec 06 08:29:10 crc kubenswrapper[4945]: I1206 08:29:10.389108 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" event={"ID":"e6cd545a-16b7-49ca-bcb9-3a733c7f4499","Type":"ContainerStarted","Data":"b04355382e4c51f8c1ce2fdb9abbdfe28af4abb8e93ebbb92465e073840d64af"} Dec 06 08:29:11 crc kubenswrapper[4945]: I1206 08:29:11.065325 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67bcf67b9f-fln2w" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.92:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.92:8080: connect: connection refused" Dec 06 08:29:11 crc kubenswrapper[4945]: I1206 08:29:11.065460 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:29:11 crc kubenswrapper[4945]: I1206 08:29:11.399759 4945 generic.go:334] "Generic (PLEG): container finished" podID="e6cd545a-16b7-49ca-bcb9-3a733c7f4499" containerID="86fd029216c1139b1419f3f590e38c02c3888d9a8eead7e3b44ff4fb4d7fb088" exitCode=0 Dec 06 08:29:11 crc kubenswrapper[4945]: I1206 08:29:11.399874 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" event={"ID":"e6cd545a-16b7-49ca-bcb9-3a733c7f4499","Type":"ContainerDied","Data":"86fd029216c1139b1419f3f590e38c02c3888d9a8eead7e3b44ff4fb4d7fb088"} Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.420569 4945 generic.go:334] "Generic (PLEG): container finished" podID="7547017f-a981-4309-979f-6776230b2198" containerID="e490063ebf7044297abdd8d76cff3a7686ad4ed811ca3a265e5ef91372cfc76b" exitCode=137 Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.421068 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bcf67b9f-fln2w" event={"ID":"7547017f-a981-4309-979f-6776230b2198","Type":"ContainerDied","Data":"e490063ebf7044297abdd8d76cff3a7686ad4ed811ca3a265e5ef91372cfc76b"} Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.425895 4945 generic.go:334] "Generic (PLEG): container finished" podID="e6cd545a-16b7-49ca-bcb9-3a733c7f4499" containerID="7f670e5f6edead6f2a4ae09de56ab81653806f0e54087b2d3f1e64148ae47601" exitCode=0 Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.425944 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" event={"ID":"e6cd545a-16b7-49ca-bcb9-3a733c7f4499","Type":"ContainerDied","Data":"7f670e5f6edead6f2a4ae09de56ab81653806f0e54087b2d3f1e64148ae47601"} Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.619044 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.781253 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7547017f-a981-4309-979f-6776230b2198-logs\") pod \"7547017f-a981-4309-979f-6776230b2198\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.781671 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-scripts\") pod \"7547017f-a981-4309-979f-6776230b2198\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.781758 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7547017f-a981-4309-979f-6776230b2198-logs" (OuterVolumeSpecName: "logs") pod "7547017f-a981-4309-979f-6776230b2198" (UID: "7547017f-a981-4309-979f-6776230b2198"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.781788 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-config-data\") pod \"7547017f-a981-4309-979f-6776230b2198\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.781829 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7547017f-a981-4309-979f-6776230b2198-horizon-secret-key\") pod \"7547017f-a981-4309-979f-6776230b2198\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.781852 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5g5b\" (UniqueName: \"kubernetes.io/projected/7547017f-a981-4309-979f-6776230b2198-kube-api-access-h5g5b\") pod \"7547017f-a981-4309-979f-6776230b2198\" (UID: \"7547017f-a981-4309-979f-6776230b2198\") " Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.782250 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7547017f-a981-4309-979f-6776230b2198-logs\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.788504 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7547017f-a981-4309-979f-6776230b2198-kube-api-access-h5g5b" (OuterVolumeSpecName: "kube-api-access-h5g5b") pod "7547017f-a981-4309-979f-6776230b2198" (UID: "7547017f-a981-4309-979f-6776230b2198"). InnerVolumeSpecName "kube-api-access-h5g5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.792372 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7547017f-a981-4309-979f-6776230b2198-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7547017f-a981-4309-979f-6776230b2198" (UID: "7547017f-a981-4309-979f-6776230b2198"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.814169 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-scripts" (OuterVolumeSpecName: "scripts") pod "7547017f-a981-4309-979f-6776230b2198" (UID: "7547017f-a981-4309-979f-6776230b2198"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.826906 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-config-data" (OuterVolumeSpecName: "config-data") pod "7547017f-a981-4309-979f-6776230b2198" (UID: "7547017f-a981-4309-979f-6776230b2198"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.883501 4945 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7547017f-a981-4309-979f-6776230b2198-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.883532 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5g5b\" (UniqueName: \"kubernetes.io/projected/7547017f-a981-4309-979f-6776230b2198-kube-api-access-h5g5b\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.883547 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:13 crc kubenswrapper[4945]: I1206 08:29:13.883580 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7547017f-a981-4309-979f-6776230b2198-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:14 crc kubenswrapper[4945]: I1206 08:29:14.441507 4945 generic.go:334] "Generic (PLEG): container finished" podID="e6cd545a-16b7-49ca-bcb9-3a733c7f4499" containerID="95220508a7fe02ca6086574578d5fb125db62d2e54aa397e3b11e15eceb5905a" exitCode=0 Dec 06 08:29:14 crc kubenswrapper[4945]: I1206 08:29:14.441615 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" event={"ID":"e6cd545a-16b7-49ca-bcb9-3a733c7f4499","Type":"ContainerDied","Data":"95220508a7fe02ca6086574578d5fb125db62d2e54aa397e3b11e15eceb5905a"} Dec 06 08:29:14 crc kubenswrapper[4945]: I1206 08:29:14.444435 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bcf67b9f-fln2w" event={"ID":"7547017f-a981-4309-979f-6776230b2198","Type":"ContainerDied","Data":"2cd687e4863598523474dcc1ed986c4a6e41f7b317f5ffa4835bb8675d2ae822"} Dec 06 08:29:14 crc kubenswrapper[4945]: I1206 08:29:14.444492 4945 scope.go:117] "RemoveContainer" containerID="73f4abd92b9c37ed61a7ad565c7e25a5ca8dd2322dc464962eafe564a1fd020e" Dec 06 08:29:14 crc kubenswrapper[4945]: I1206 08:29:14.444519 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67bcf67b9f-fln2w" Dec 06 08:29:14 crc kubenswrapper[4945]: I1206 08:29:14.495329 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67bcf67b9f-fln2w"] Dec 06 08:29:14 crc kubenswrapper[4945]: I1206 08:29:14.506557 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-67bcf67b9f-fln2w"] Dec 06 08:29:14 crc kubenswrapper[4945]: I1206 08:29:14.628936 4945 scope.go:117] "RemoveContainer" containerID="e490063ebf7044297abdd8d76cff3a7686ad4ed811ca3a265e5ef91372cfc76b" Dec 06 08:29:14 crc kubenswrapper[4945]: I1206 08:29:14.965356 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7547017f-a981-4309-979f-6776230b2198" path="/var/lib/kubelet/pods/7547017f-a981-4309-979f-6776230b2198/volumes" Dec 06 08:29:15 crc kubenswrapper[4945]: I1206 08:29:15.824433 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.025345 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-util\") pod \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.025699 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl8l7\" (UniqueName: \"kubernetes.io/projected/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-kube-api-access-jl8l7\") pod \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.025719 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-bundle\") pod \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\" (UID: \"e6cd545a-16b7-49ca-bcb9-3a733c7f4499\") " Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.028125 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-bundle" (OuterVolumeSpecName: "bundle") pod "e6cd545a-16b7-49ca-bcb9-3a733c7f4499" (UID: "e6cd545a-16b7-49ca-bcb9-3a733c7f4499"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.032050 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-kube-api-access-jl8l7" (OuterVolumeSpecName: "kube-api-access-jl8l7") pod "e6cd545a-16b7-49ca-bcb9-3a733c7f4499" (UID: "e6cd545a-16b7-49ca-bcb9-3a733c7f4499"). InnerVolumeSpecName "kube-api-access-jl8l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.039834 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-util" (OuterVolumeSpecName: "util") pod "e6cd545a-16b7-49ca-bcb9-3a733c7f4499" (UID: "e6cd545a-16b7-49ca-bcb9-3a733c7f4499"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.056082 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7c24-account-create-update-cgl54"] Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.066200 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7c24-account-create-update-cgl54"] Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.074894 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-tpdn5"] Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.083482 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-tpdn5"] Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.127608 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-util\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.127638 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl8l7\" (UniqueName: \"kubernetes.io/projected/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-kube-api-access-jl8l7\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.127647 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6cd545a-16b7-49ca-bcb9-3a733c7f4499-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.468061 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" event={"ID":"e6cd545a-16b7-49ca-bcb9-3a733c7f4499","Type":"ContainerDied","Data":"b04355382e4c51f8c1ce2fdb9abbdfe28af4abb8e93ebbb92465e073840d64af"} Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.468110 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b04355382e4c51f8c1ce2fdb9abbdfe28af4abb8e93ebbb92465e073840d64af" Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.468146 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl" Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.986422 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a1b76d4-f78c-4972-832c-c586920f6a0f" path="/var/lib/kubelet/pods/5a1b76d4-f78c-4972-832c-c586920f6a0f/volumes" Dec 06 08:29:16 crc kubenswrapper[4945]: I1206 08:29:16.989972 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9a17984-24af-49d1-ad27-54655f405ee8" path="/var/lib/kubelet/pods/e9a17984-24af-49d1-ad27-54655f405ee8/volumes" Dec 06 08:29:25 crc kubenswrapper[4945]: I1206 08:29:25.039803 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-g85nz"] Dec 06 08:29:25 crc kubenswrapper[4945]: I1206 08:29:25.048206 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-g85nz"] Dec 06 08:29:25 crc kubenswrapper[4945]: I1206 08:29:25.921887 4945 scope.go:117] "RemoveContainer" containerID="0362c706459ad07def527950a4c7330ed2c8ae1142960517ac2d91c820c4ec43" Dec 06 08:29:25 crc kubenswrapper[4945]: I1206 08:29:25.951308 4945 scope.go:117] "RemoveContainer" containerID="fd3b89d05d6b2916c2c89ae1e0a0999380306c5d4791b2531a52dc69babb7ade" Dec 06 08:29:26 crc kubenswrapper[4945]: I1206 08:29:26.014188 4945 scope.go:117] "RemoveContainer" containerID="05f147cc3615e396cd747214bc8f0f527438fdca0ae3ee4612a8f1fedfc3b151" Dec 06 08:29:26 crc kubenswrapper[4945]: I1206 08:29:26.061919 4945 scope.go:117] "RemoveContainer" containerID="8bc9b2b9b06e8a3035e1111ea4fed8b6559b28e9f717a825bad06b2c07582dda" Dec 06 08:29:26 crc kubenswrapper[4945]: I1206 08:29:26.102046 4945 scope.go:117] "RemoveContainer" containerID="4c015654cc5d115f268eeeb4b314d7df0e9c0f967efd273d9f4d15c824cc6864" Dec 06 08:29:26 crc kubenswrapper[4945]: I1206 08:29:26.981766 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="231b069e-3731-42fa-9431-1b9081ea9d0d" path="/var/lib/kubelet/pods/231b069e-3731-42fa-9431-1b9081ea9d0d/volumes" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.032928 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv"] Dec 06 08:29:28 crc kubenswrapper[4945]: E1206 08:29:28.035554 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon-log" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.035581 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon-log" Dec 06 08:29:28 crc kubenswrapper[4945]: E1206 08:29:28.035601 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6cd545a-16b7-49ca-bcb9-3a733c7f4499" containerName="pull" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.035606 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6cd545a-16b7-49ca-bcb9-3a733c7f4499" containerName="pull" Dec 06 08:29:28 crc kubenswrapper[4945]: E1206 08:29:28.035622 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6cd545a-16b7-49ca-bcb9-3a733c7f4499" containerName="extract" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.035627 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6cd545a-16b7-49ca-bcb9-3a733c7f4499" containerName="extract" Dec 06 08:29:28 crc kubenswrapper[4945]: E1206 08:29:28.035651 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6cd545a-16b7-49ca-bcb9-3a733c7f4499" containerName="util" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.035657 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6cd545a-16b7-49ca-bcb9-3a733c7f4499" containerName="util" Dec 06 08:29:28 crc kubenswrapper[4945]: E1206 08:29:28.035670 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.035676 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.035861 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6cd545a-16b7-49ca-bcb9-3a733c7f4499" containerName="extract" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.035873 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.035889 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7547017f-a981-4309-979f-6776230b2198" containerName="horizon-log" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.036593 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.044487 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.044592 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-89rvz" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.044632 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.050097 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sgwd\" (UniqueName: \"kubernetes.io/projected/37a5d5bb-936f-4c5b-a808-e88fc434aa2f-kube-api-access-6sgwd\") pod \"obo-prometheus-operator-668cf9dfbb-kmwcv\" (UID: \"37a5d5bb-936f-4c5b-a808-e88fc434aa2f\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.057427 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv"] Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.085622 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k"] Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.086971 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.094016 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-m5d67" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.098777 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.109023 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz"] Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.110557 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.123424 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k"] Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.136564 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz"] Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.152740 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/11477f2b-5358-45d2-921c-d2ced41fe1f0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-c5crz\" (UID: \"11477f2b-5358-45d2-921c-d2ced41fe1f0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.153094 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/11477f2b-5358-45d2-921c-d2ced41fe1f0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-c5crz\" (UID: \"11477f2b-5358-45d2-921c-d2ced41fe1f0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.153288 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sgwd\" (UniqueName: \"kubernetes.io/projected/37a5d5bb-936f-4c5b-a808-e88fc434aa2f-kube-api-access-6sgwd\") pod \"obo-prometheus-operator-668cf9dfbb-kmwcv\" (UID: \"37a5d5bb-936f-4c5b-a808-e88fc434aa2f\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.153385 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19169e62-1838-4277-a99d-f9b2944e37c2-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k\" (UID: \"19169e62-1838-4277-a99d-f9b2944e37c2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.153463 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19169e62-1838-4277-a99d-f9b2944e37c2-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k\" (UID: \"19169e62-1838-4277-a99d-f9b2944e37c2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.186129 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sgwd\" (UniqueName: \"kubernetes.io/projected/37a5d5bb-936f-4c5b-a808-e88fc434aa2f-kube-api-access-6sgwd\") pod \"obo-prometheus-operator-668cf9dfbb-kmwcv\" (UID: \"37a5d5bb-936f-4c5b-a808-e88fc434aa2f\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.255452 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/11477f2b-5358-45d2-921c-d2ced41fe1f0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-c5crz\" (UID: \"11477f2b-5358-45d2-921c-d2ced41fe1f0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.255536 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/11477f2b-5358-45d2-921c-d2ced41fe1f0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-c5crz\" (UID: \"11477f2b-5358-45d2-921c-d2ced41fe1f0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.255566 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19169e62-1838-4277-a99d-f9b2944e37c2-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k\" (UID: \"19169e62-1838-4277-a99d-f9b2944e37c2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.255585 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19169e62-1838-4277-a99d-f9b2944e37c2-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k\" (UID: \"19169e62-1838-4277-a99d-f9b2944e37c2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.261667 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-twxbc"] Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.263912 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/11477f2b-5358-45d2-921c-d2ced41fe1f0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-c5crz\" (UID: \"11477f2b-5358-45d2-921c-d2ced41fe1f0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.266142 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19169e62-1838-4277-a99d-f9b2944e37c2-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k\" (UID: \"19169e62-1838-4277-a99d-f9b2944e37c2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.269268 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19169e62-1838-4277-a99d-f9b2944e37c2-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k\" (UID: \"19169e62-1838-4277-a99d-f9b2944e37c2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.270115 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.276904 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/11477f2b-5358-45d2-921c-d2ced41fe1f0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-546584f4f9-c5crz\" (UID: \"11477f2b-5358-45d2-921c-d2ced41fe1f0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.278695 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.278882 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-wwcnj" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.306591 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-twxbc"] Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.357632 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/ccaddde6-c87e-4aba-b71f-4974135fa01a-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-twxbc\" (UID: \"ccaddde6-c87e-4aba-b71f-4974135fa01a\") " pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.357720 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nv4k\" (UniqueName: \"kubernetes.io/projected/ccaddde6-c87e-4aba-b71f-4974135fa01a-kube-api-access-5nv4k\") pod \"observability-operator-d8bb48f5d-twxbc\" (UID: \"ccaddde6-c87e-4aba-b71f-4974135fa01a\") " pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.373365 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.409177 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.443710 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.459213 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/ccaddde6-c87e-4aba-b71f-4974135fa01a-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-twxbc\" (UID: \"ccaddde6-c87e-4aba-b71f-4974135fa01a\") " pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.459336 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nv4k\" (UniqueName: \"kubernetes.io/projected/ccaddde6-c87e-4aba-b71f-4974135fa01a-kube-api-access-5nv4k\") pod \"observability-operator-d8bb48f5d-twxbc\" (UID: \"ccaddde6-c87e-4aba-b71f-4974135fa01a\") " pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.464040 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-mg69q"] Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.465323 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-mg69q" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.467751 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-rlswr" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.477059 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/ccaddde6-c87e-4aba-b71f-4974135fa01a-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-twxbc\" (UID: \"ccaddde6-c87e-4aba-b71f-4974135fa01a\") " pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.483978 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nv4k\" (UniqueName: \"kubernetes.io/projected/ccaddde6-c87e-4aba-b71f-4974135fa01a-kube-api-access-5nv4k\") pod \"observability-operator-d8bb48f5d-twxbc\" (UID: \"ccaddde6-c87e-4aba-b71f-4974135fa01a\") " pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.508355 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-mg69q"] Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.668399 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e66cba6-f587-4390-9f2c-69003f179cae-openshift-service-ca\") pod \"perses-operator-5446b9c989-mg69q\" (UID: \"1e66cba6-f587-4390-9f2c-69003f179cae\") " pod="openshift-operators/perses-operator-5446b9c989-mg69q" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.668751 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpv7b\" (UniqueName: \"kubernetes.io/projected/1e66cba6-f587-4390-9f2c-69003f179cae-kube-api-access-lpv7b\") pod \"perses-operator-5446b9c989-mg69q\" (UID: \"1e66cba6-f587-4390-9f2c-69003f179cae\") " pod="openshift-operators/perses-operator-5446b9c989-mg69q" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.708718 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.770522 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e66cba6-f587-4390-9f2c-69003f179cae-openshift-service-ca\") pod \"perses-operator-5446b9c989-mg69q\" (UID: \"1e66cba6-f587-4390-9f2c-69003f179cae\") " pod="openshift-operators/perses-operator-5446b9c989-mg69q" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.770566 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpv7b\" (UniqueName: \"kubernetes.io/projected/1e66cba6-f587-4390-9f2c-69003f179cae-kube-api-access-lpv7b\") pod \"perses-operator-5446b9c989-mg69q\" (UID: \"1e66cba6-f587-4390-9f2c-69003f179cae\") " pod="openshift-operators/perses-operator-5446b9c989-mg69q" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.771673 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e66cba6-f587-4390-9f2c-69003f179cae-openshift-service-ca\") pod \"perses-operator-5446b9c989-mg69q\" (UID: \"1e66cba6-f587-4390-9f2c-69003f179cae\") " pod="openshift-operators/perses-operator-5446b9c989-mg69q" Dec 06 08:29:28 crc kubenswrapper[4945]: I1206 08:29:28.807324 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpv7b\" (UniqueName: \"kubernetes.io/projected/1e66cba6-f587-4390-9f2c-69003f179cae-kube-api-access-lpv7b\") pod \"perses-operator-5446b9c989-mg69q\" (UID: \"1e66cba6-f587-4390-9f2c-69003f179cae\") " pod="openshift-operators/perses-operator-5446b9c989-mg69q" Dec 06 08:29:29 crc kubenswrapper[4945]: I1206 08:29:29.105296 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-mg69q" Dec 06 08:29:29 crc kubenswrapper[4945]: I1206 08:29:29.115387 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k"] Dec 06 08:29:29 crc kubenswrapper[4945]: W1206 08:29:29.130803 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19169e62_1838_4277_a99d_f9b2944e37c2.slice/crio-3e9f7172f69795204bbf57ec894da847631cf723baae1a35b3f50356ecc42afb WatchSource:0}: Error finding container 3e9f7172f69795204bbf57ec894da847631cf723baae1a35b3f50356ecc42afb: Status 404 returned error can't find the container with id 3e9f7172f69795204bbf57ec894da847631cf723baae1a35b3f50356ecc42afb Dec 06 08:29:29 crc kubenswrapper[4945]: I1206 08:29:29.145448 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv"] Dec 06 08:29:29 crc kubenswrapper[4945]: I1206 08:29:29.383766 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-twxbc"] Dec 06 08:29:29 crc kubenswrapper[4945]: I1206 08:29:29.416387 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz"] Dec 06 08:29:29 crc kubenswrapper[4945]: I1206 08:29:29.595858 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv" event={"ID":"37a5d5bb-936f-4c5b-a808-e88fc434aa2f","Type":"ContainerStarted","Data":"327f5916ea98f58960a74a12f8aaa11447b211666cf264c95ef7106c5f56743f"} Dec 06 08:29:29 crc kubenswrapper[4945]: I1206 08:29:29.598055 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" event={"ID":"11477f2b-5358-45d2-921c-d2ced41fe1f0","Type":"ContainerStarted","Data":"d7a5b516c9457512715312a8592bb20eb5f5a13855633ec9d3ee19a6889210b5"} Dec 06 08:29:29 crc kubenswrapper[4945]: I1206 08:29:29.599921 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" event={"ID":"ccaddde6-c87e-4aba-b71f-4974135fa01a","Type":"ContainerStarted","Data":"3281a8ce95654923eecb6ec7758bc275688e2d6f6da76397d7c61ae4242e008f"} Dec 06 08:29:29 crc kubenswrapper[4945]: I1206 08:29:29.601658 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" event={"ID":"19169e62-1838-4277-a99d-f9b2944e37c2","Type":"ContainerStarted","Data":"3e9f7172f69795204bbf57ec894da847631cf723baae1a35b3f50356ecc42afb"} Dec 06 08:29:29 crc kubenswrapper[4945]: I1206 08:29:29.609725 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-mg69q"] Dec 06 08:29:30 crc kubenswrapper[4945]: I1206 08:29:30.632486 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-mg69q" event={"ID":"1e66cba6-f587-4390-9f2c-69003f179cae","Type":"ContainerStarted","Data":"6b409eb46a87f3b044d6229675b7e79413c480b98e37ac476c2360ca52877f49"} Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.726498 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" event={"ID":"11477f2b-5358-45d2-921c-d2ced41fe1f0","Type":"ContainerStarted","Data":"001ccd73cb1aefb12432ef50ff87370db1debd116170fac6cf1e82c9076bc91f"} Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.729268 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" event={"ID":"ccaddde6-c87e-4aba-b71f-4974135fa01a","Type":"ContainerStarted","Data":"93cf5884be046da5a9bdd04914215f6627ea89f021eaab6d1bdbf16d0c32253b"} Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.729476 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.731877 4945 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-twxbc container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.106:8081/healthz\": dial tcp 10.217.1.106:8081: connect: connection refused" start-of-body= Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.731920 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" podUID="ccaddde6-c87e-4aba-b71f-4974135fa01a" containerName="operator" probeResult="failure" output="Get \"http://10.217.1.106:8081/healthz\": dial tcp 10.217.1.106:8081: connect: connection refused" Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.732063 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-mg69q" event={"ID":"1e66cba6-f587-4390-9f2c-69003f179cae","Type":"ContainerStarted","Data":"fd24d2277d152a9a6e6d0488db0f6cbe5fceaf79bb3fc183d7b8abe30ca6261b"} Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.732178 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-mg69q" Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.734358 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" event={"ID":"19169e62-1838-4277-a99d-f9b2944e37c2","Type":"ContainerStarted","Data":"8baed89813d59062a18ba2afd187b703c76d089cd559083f8652d9606cb28636"} Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.747787 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-c5crz" podStartSLOduration=2.122782079 podStartE2EDuration="10.747771529s" podCreationTimestamp="2025-12-06 08:29:28 +0000 UTC" firstStartedPulling="2025-12-06 08:29:29.43916587 +0000 UTC m=+5782.894026914" lastFinishedPulling="2025-12-06 08:29:38.06415532 +0000 UTC m=+5791.519016364" observedRunningTime="2025-12-06 08:29:38.745666432 +0000 UTC m=+5792.200527486" watchObservedRunningTime="2025-12-06 08:29:38.747771529 +0000 UTC m=+5792.202632563" Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.778774 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-mg69q" podStartSLOduration=2.326937652 podStartE2EDuration="10.778752778s" podCreationTimestamp="2025-12-06 08:29:28 +0000 UTC" firstStartedPulling="2025-12-06 08:29:29.615509748 +0000 UTC m=+5783.070370802" lastFinishedPulling="2025-12-06 08:29:38.067324884 +0000 UTC m=+5791.522185928" observedRunningTime="2025-12-06 08:29:38.774222436 +0000 UTC m=+5792.229083481" watchObservedRunningTime="2025-12-06 08:29:38.778752778 +0000 UTC m=+5792.233613822" Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.795388 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.795440 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.795486 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.796415 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"170cc98b8d9a5021f285e8b2c4ca1f3a5177f12dc09aebb616fcafb0121da204"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.796498 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://170cc98b8d9a5021f285e8b2c4ca1f3a5177f12dc09aebb616fcafb0121da204" gracePeriod=600 Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.818931 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" podStartSLOduration=2.044975827 podStartE2EDuration="10.818913652s" podCreationTimestamp="2025-12-06 08:29:28 +0000 UTC" firstStartedPulling="2025-12-06 08:29:29.417445259 +0000 UTC m=+5782.872306303" lastFinishedPulling="2025-12-06 08:29:38.191383084 +0000 UTC m=+5791.646244128" observedRunningTime="2025-12-06 08:29:38.811073162 +0000 UTC m=+5792.265934206" watchObservedRunningTime="2025-12-06 08:29:38.818913652 +0000 UTC m=+5792.273774686" Dec 06 08:29:38 crc kubenswrapper[4945]: I1206 08:29:38.841136 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k" podStartSLOduration=1.91052054 podStartE2EDuration="10.841115386s" podCreationTimestamp="2025-12-06 08:29:28 +0000 UTC" firstStartedPulling="2025-12-06 08:29:29.133904123 +0000 UTC m=+5782.588765167" lastFinishedPulling="2025-12-06 08:29:38.064498969 +0000 UTC m=+5791.519360013" observedRunningTime="2025-12-06 08:29:38.836353789 +0000 UTC m=+5792.291214843" watchObservedRunningTime="2025-12-06 08:29:38.841115386 +0000 UTC m=+5792.295976440" Dec 06 08:29:39 crc kubenswrapper[4945]: I1206 08:29:39.747112 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv" event={"ID":"37a5d5bb-936f-4c5b-a808-e88fc434aa2f","Type":"ContainerStarted","Data":"06e9c845c2bdd224481ed5d70303d7a1d7b4cb3f63839010fd1b9f15fc9593ec"} Dec 06 08:29:39 crc kubenswrapper[4945]: I1206 08:29:39.756078 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="170cc98b8d9a5021f285e8b2c4ca1f3a5177f12dc09aebb616fcafb0121da204" exitCode=0 Dec 06 08:29:39 crc kubenswrapper[4945]: I1206 08:29:39.756252 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"170cc98b8d9a5021f285e8b2c4ca1f3a5177f12dc09aebb616fcafb0121da204"} Dec 06 08:29:39 crc kubenswrapper[4945]: I1206 08:29:39.756324 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027"} Dec 06 08:29:39 crc kubenswrapper[4945]: I1206 08:29:39.756351 4945 scope.go:117] "RemoveContainer" containerID="d56349f61c62859aef63cca7a0d1a18a3cc4d990d6fd401fe4df0b4a02d3cc39" Dec 06 08:29:39 crc kubenswrapper[4945]: I1206 08:29:39.760123 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-twxbc" Dec 06 08:29:39 crc kubenswrapper[4945]: I1206 08:29:39.775928 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kmwcv" podStartSLOduration=2.791282915 podStartE2EDuration="11.775904637s" podCreationTimestamp="2025-12-06 08:29:28 +0000 UTC" firstStartedPulling="2025-12-06 08:29:29.172600708 +0000 UTC m=+5782.627461752" lastFinishedPulling="2025-12-06 08:29:38.15722243 +0000 UTC m=+5791.612083474" observedRunningTime="2025-12-06 08:29:39.768860678 +0000 UTC m=+5793.223721722" watchObservedRunningTime="2025-12-06 08:29:39.775904637 +0000 UTC m=+5793.230765691" Dec 06 08:29:49 crc kubenswrapper[4945]: I1206 08:29:49.109416 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-mg69q" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.590027 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kfdll"] Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.592812 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.606506 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kfdll"] Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.720973 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-utilities\") pod \"redhat-marketplace-kfdll\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.721014 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-catalog-content\") pod \"redhat-marketplace-kfdll\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.721055 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8w7s\" (UniqueName: \"kubernetes.io/projected/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-kube-api-access-s8w7s\") pod \"redhat-marketplace-kfdll\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.822651 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-utilities\") pod \"redhat-marketplace-kfdll\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.823041 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-utilities\") pod \"redhat-marketplace-kfdll\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.823105 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-catalog-content\") pod \"redhat-marketplace-kfdll\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.822765 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-catalog-content\") pod \"redhat-marketplace-kfdll\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.823199 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8w7s\" (UniqueName: \"kubernetes.io/projected/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-kube-api-access-s8w7s\") pod \"redhat-marketplace-kfdll\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.875352 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8w7s\" (UniqueName: \"kubernetes.io/projected/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-kube-api-access-s8w7s\") pod \"redhat-marketplace-kfdll\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:51 crc kubenswrapper[4945]: I1206 08:29:51.915737 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.368338 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.368842 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="039bb79d-fba0-413e-844f-e21e8ec625d6" containerName="openstackclient" containerID="cri-o://8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61" gracePeriod=2 Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.378991 4945 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="039bb79d-fba0-413e-844f-e21e8ec625d6" podUID="d7b129e8-fa9c-47f4-bb42-ed7867d79609" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.412703 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.426363 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 06 08:29:52 crc kubenswrapper[4945]: E1206 08:29:52.428986 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="039bb79d-fba0-413e-844f-e21e8ec625d6" containerName="openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.429018 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="039bb79d-fba0-413e-844f-e21e8ec625d6" containerName="openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.429269 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="039bb79d-fba0-413e-844f-e21e8ec625d6" containerName="openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.430924 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.439082 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.540764 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d7b129e8-fa9c-47f4-bb42-ed7867d79609-openstack-config\") pod \"openstackclient\" (UID: \"d7b129e8-fa9c-47f4-bb42-ed7867d79609\") " pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.540820 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d7b129e8-fa9c-47f4-bb42-ed7867d79609-openstack-config-secret\") pod \"openstackclient\" (UID: \"d7b129e8-fa9c-47f4-bb42-ed7867d79609\") " pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.540913 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4ptl\" (UniqueName: \"kubernetes.io/projected/d7b129e8-fa9c-47f4-bb42-ed7867d79609-kube-api-access-k4ptl\") pod \"openstackclient\" (UID: \"d7b129e8-fa9c-47f4-bb42-ed7867d79609\") " pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.642316 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d7b129e8-fa9c-47f4-bb42-ed7867d79609-openstack-config-secret\") pod \"openstackclient\" (UID: \"d7b129e8-fa9c-47f4-bb42-ed7867d79609\") " pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.642421 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4ptl\" (UniqueName: \"kubernetes.io/projected/d7b129e8-fa9c-47f4-bb42-ed7867d79609-kube-api-access-k4ptl\") pod \"openstackclient\" (UID: \"d7b129e8-fa9c-47f4-bb42-ed7867d79609\") " pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.642508 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d7b129e8-fa9c-47f4-bb42-ed7867d79609-openstack-config\") pod \"openstackclient\" (UID: \"d7b129e8-fa9c-47f4-bb42-ed7867d79609\") " pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.643483 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d7b129e8-fa9c-47f4-bb42-ed7867d79609-openstack-config\") pod \"openstackclient\" (UID: \"d7b129e8-fa9c-47f4-bb42-ed7867d79609\") " pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.660610 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d7b129e8-fa9c-47f4-bb42-ed7867d79609-openstack-config-secret\") pod \"openstackclient\" (UID: \"d7b129e8-fa9c-47f4-bb42-ed7867d79609\") " pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.690937 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.696211 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.702322 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-n6kq6" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.725772 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.726944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4ptl\" (UniqueName: \"kubernetes.io/projected/d7b129e8-fa9c-47f4-bb42-ed7867d79609-kube-api-access-k4ptl\") pod \"openstackclient\" (UID: \"d7b129e8-fa9c-47f4-bb42-ed7867d79609\") " pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.774315 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.851322 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-996l9\" (UniqueName: \"kubernetes.io/projected/2e198f48-5c2b-4834-9c95-934480712bf2-kube-api-access-996l9\") pod \"kube-state-metrics-0\" (UID: \"2e198f48-5c2b-4834-9c95-934480712bf2\") " pod="openstack/kube-state-metrics-0" Dec 06 08:29:52 crc kubenswrapper[4945]: I1206 08:29:52.952682 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-996l9\" (UniqueName: \"kubernetes.io/projected/2e198f48-5c2b-4834-9c95-934480712bf2-kube-api-access-996l9\") pod \"kube-state-metrics-0\" (UID: \"2e198f48-5c2b-4834-9c95-934480712bf2\") " pod="openstack/kube-state-metrics-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.004610 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-996l9\" (UniqueName: \"kubernetes.io/projected/2e198f48-5c2b-4834-9c95-934480712bf2-kube-api-access-996l9\") pod \"kube-state-metrics-0\" (UID: \"2e198f48-5c2b-4834-9c95-934480712bf2\") " pod="openstack/kube-state-metrics-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.038165 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kfdll"] Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.114317 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.571241 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.577844 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.584965 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.585201 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-7d2mv" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.589522 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.589691 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.601027 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.646417 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.697159 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.697308 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5xtq\" (UniqueName: \"kubernetes.io/projected/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-kube-api-access-f5xtq\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.697342 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.697408 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.697468 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.697493 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.697509 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.805359 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.805430 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.805459 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.805477 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.805529 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.805597 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5xtq\" (UniqueName: \"kubernetes.io/projected/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-kube-api-access-f5xtq\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.805619 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.812056 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.815047 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.818892 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.829111 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.830782 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.841068 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.853571 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5xtq\" (UniqueName: \"kubernetes.io/projected/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-kube-api-access-f5xtq\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.870852 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/5f1d750f-7e4b-487c-a552-5bf7c183c7ad-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5f1d750f-7e4b-487c-a552-5bf7c183c7ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.923205 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 06 08:29:53 crc kubenswrapper[4945]: I1206 08:29:53.933980 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfdll" event={"ID":"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d","Type":"ContainerStarted","Data":"2a3a7f7b80ad86ccdbab3dbe71136ebd61c47f01cfb0a599ad4d1ada1360dd8a"} Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.137259 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.145921 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.157729 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.157827 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.158506 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-dvd7t" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.159155 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.159644 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.165042 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.169314 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.307268 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.323374 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4c654f23-0692-4d36-918c-3e669d026c44-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.323417 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c654f23-0692-4d36-918c-3e669d026c44-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.323445 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b10c0a7b-40d4-44c4-8323-9fb23818051d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b10c0a7b-40d4-44c4-8323-9fb23818051d\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.323643 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4c654f23-0692-4d36-918c-3e669d026c44-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.323765 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c654f23-0692-4d36-918c-3e669d026c44-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.323807 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c654f23-0692-4d36-918c-3e669d026c44-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.323855 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8hkq\" (UniqueName: \"kubernetes.io/projected/4c654f23-0692-4d36-918c-3e669d026c44-kube-api-access-z8hkq\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.324083 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4c654f23-0692-4d36-918c-3e669d026c44-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: W1206 08:29:54.345710 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e198f48_5c2b_4834_9c95_934480712bf2.slice/crio-5f10f390bb5e3842792f6dfd105bfc28278e63f7b237059ca902c835396eeec8 WatchSource:0}: Error finding container 5f10f390bb5e3842792f6dfd105bfc28278e63f7b237059ca902c835396eeec8: Status 404 returned error can't find the container with id 5f10f390bb5e3842792f6dfd105bfc28278e63f7b237059ca902c835396eeec8 Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.428110 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4c654f23-0692-4d36-918c-3e669d026c44-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.428526 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4c654f23-0692-4d36-918c-3e669d026c44-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.428550 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c654f23-0692-4d36-918c-3e669d026c44-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.428574 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b10c0a7b-40d4-44c4-8323-9fb23818051d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b10c0a7b-40d4-44c4-8323-9fb23818051d\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.428632 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4c654f23-0692-4d36-918c-3e669d026c44-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.428662 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c654f23-0692-4d36-918c-3e669d026c44-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.428682 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c654f23-0692-4d36-918c-3e669d026c44-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.428702 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8hkq\" (UniqueName: \"kubernetes.io/projected/4c654f23-0692-4d36-918c-3e669d026c44-kube-api-access-z8hkq\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.430558 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4c654f23-0692-4d36-918c-3e669d026c44-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.434221 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c654f23-0692-4d36-918c-3e669d026c44-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.441109 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.441162 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b10c0a7b-40d4-44c4-8323-9fb23818051d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b10c0a7b-40d4-44c4-8323-9fb23818051d\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/56be9ce0ff92af58e434050b11717cf77d7b40d61afbebd591afcb498bd8cbf0/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.444062 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4c654f23-0692-4d36-918c-3e669d026c44-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.448766 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c654f23-0692-4d36-918c-3e669d026c44-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.450718 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4c654f23-0692-4d36-918c-3e669d026c44-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.457880 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8hkq\" (UniqueName: \"kubernetes.io/projected/4c654f23-0692-4d36-918c-3e669d026c44-kube-api-access-z8hkq\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.461211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c654f23-0692-4d36-918c-3e669d026c44-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.590133 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b10c0a7b-40d4-44c4-8323-9fb23818051d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b10c0a7b-40d4-44c4-8323-9fb23818051d\") pod \"prometheus-metric-storage-0\" (UID: \"4c654f23-0692-4d36-918c-3e669d026c44\") " pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.632798 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.730978 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.840873 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config\") pod \"039bb79d-fba0-413e-844f-e21e8ec625d6\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.841027 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config-secret\") pod \"039bb79d-fba0-413e-844f-e21e8ec625d6\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.841083 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bclr7\" (UniqueName: \"kubernetes.io/projected/039bb79d-fba0-413e-844f-e21e8ec625d6-kube-api-access-bclr7\") pod \"039bb79d-fba0-413e-844f-e21e8ec625d6\" (UID: \"039bb79d-fba0-413e-844f-e21e8ec625d6\") " Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.848069 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/039bb79d-fba0-413e-844f-e21e8ec625d6-kube-api-access-bclr7" (OuterVolumeSpecName: "kube-api-access-bclr7") pod "039bb79d-fba0-413e-844f-e21e8ec625d6" (UID: "039bb79d-fba0-413e-844f-e21e8ec625d6"). InnerVolumeSpecName "kube-api-access-bclr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.899525 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.915871 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "039bb79d-fba0-413e-844f-e21e8ec625d6" (UID: "039bb79d-fba0-413e-844f-e21e8ec625d6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.936635 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "039bb79d-fba0-413e-844f-e21e8ec625d6" (UID: "039bb79d-fba0-413e-844f-e21e8ec625d6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.943422 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.943455 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/039bb79d-fba0-413e-844f-e21e8ec625d6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.943477 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bclr7\" (UniqueName: \"kubernetes.io/projected/039bb79d-fba0-413e-844f-e21e8ec625d6-kube-api-access-bclr7\") on node \"crc\" DevicePath \"\"" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.975183 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="039bb79d-fba0-413e-844f-e21e8ec625d6" path="/var/lib/kubelet/pods/039bb79d-fba0-413e-844f-e21e8ec625d6/volumes" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.976829 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2e198f48-5c2b-4834-9c95-934480712bf2","Type":"ContainerStarted","Data":"5f10f390bb5e3842792f6dfd105bfc28278e63f7b237059ca902c835396eeec8"} Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.977085 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d7b129e8-fa9c-47f4-bb42-ed7867d79609","Type":"ContainerStarted","Data":"1b628f37f190a8e7aae9930e4b3c6bae6d7a57b55642d38cf87c327a300b3a04"} Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.977163 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d7b129e8-fa9c-47f4-bb42-ed7867d79609","Type":"ContainerStarted","Data":"fd49fb9dd9b2920dc56409ea27a5f012c73051228a0604e1d570419c5cfbde4c"} Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.977457 4945 generic.go:334] "Generic (PLEG): container finished" podID="039bb79d-fba0-413e-844f-e21e8ec625d6" containerID="8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61" exitCode=137 Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.977543 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.977658 4945 scope.go:117] "RemoveContainer" containerID="8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.985272 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.985257029 podStartE2EDuration="2.985257029s" podCreationTimestamp="2025-12-06 08:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:29:54.98376536 +0000 UTC m=+5808.438626404" watchObservedRunningTime="2025-12-06 08:29:54.985257029 +0000 UTC m=+5808.440118073" Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.995962 4945 generic.go:334] "Generic (PLEG): container finished" podID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerID="f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c" exitCode=0 Dec 06 08:29:54 crc kubenswrapper[4945]: I1206 08:29:54.996841 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfdll" event={"ID":"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d","Type":"ContainerDied","Data":"f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c"} Dec 06 08:29:55 crc kubenswrapper[4945]: I1206 08:29:55.009707 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f1d750f-7e4b-487c-a552-5bf7c183c7ad","Type":"ContainerStarted","Data":"2df11f280ebbdd56a4bad1c3fa99df8eda1ee9c43d2d6a4201ad8f5bb7190582"} Dec 06 08:29:55 crc kubenswrapper[4945]: I1206 08:29:55.041654 4945 scope.go:117] "RemoveContainer" containerID="8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61" Dec 06 08:29:55 crc kubenswrapper[4945]: E1206 08:29:55.043062 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61\": container with ID starting with 8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61 not found: ID does not exist" containerID="8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61" Dec 06 08:29:55 crc kubenswrapper[4945]: I1206 08:29:55.043117 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61"} err="failed to get container status \"8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61\": rpc error: code = NotFound desc = could not find container \"8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61\": container with ID starting with 8b60fdb44acd0462c15d4e4cbb2faf8fa02490ca0a5fe3df7d3294ae41993c61 not found: ID does not exist" Dec 06 08:29:55 crc kubenswrapper[4945]: W1206 08:29:55.265994 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c654f23_0692_4d36_918c_3e669d026c44.slice/crio-786e667bd79e9e8236863307c6a964d6d4fd493a01123cab535bff65f3113c0c WatchSource:0}: Error finding container 786e667bd79e9e8236863307c6a964d6d4fd493a01123cab535bff65f3113c0c: Status 404 returned error can't find the container with id 786e667bd79e9e8236863307c6a964d6d4fd493a01123cab535bff65f3113c0c Dec 06 08:29:55 crc kubenswrapper[4945]: I1206 08:29:55.268068 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 08:29:56 crc kubenswrapper[4945]: I1206 08:29:56.020777 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c654f23-0692-4d36-918c-3e669d026c44","Type":"ContainerStarted","Data":"786e667bd79e9e8236863307c6a964d6d4fd493a01123cab535bff65f3113c0c"} Dec 06 08:29:56 crc kubenswrapper[4945]: I1206 08:29:56.034206 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfdll" event={"ID":"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d","Type":"ContainerStarted","Data":"f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f"} Dec 06 08:29:56 crc kubenswrapper[4945]: I1206 08:29:56.039676 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2e198f48-5c2b-4834-9c95-934480712bf2","Type":"ContainerStarted","Data":"80c8d165f454930e9a29decb9eba82a291ee7b96de0e850135e75f38b3e1f76e"} Dec 06 08:29:56 crc kubenswrapper[4945]: I1206 08:29:56.039713 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 08:29:56 crc kubenswrapper[4945]: I1206 08:29:56.990640 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=4.606595218 podStartE2EDuration="4.990603182s" podCreationTimestamp="2025-12-06 08:29:52 +0000 UTC" firstStartedPulling="2025-12-06 08:29:54.348217746 +0000 UTC m=+5807.803078790" lastFinishedPulling="2025-12-06 08:29:54.73222571 +0000 UTC m=+5808.187086754" observedRunningTime="2025-12-06 08:29:56.072781946 +0000 UTC m=+5809.527642990" watchObservedRunningTime="2025-12-06 08:29:56.990603182 +0000 UTC m=+5810.445464226" Dec 06 08:29:57 crc kubenswrapper[4945]: I1206 08:29:57.051145 4945 generic.go:334] "Generic (PLEG): container finished" podID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerID="f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f" exitCode=0 Dec 06 08:29:57 crc kubenswrapper[4945]: I1206 08:29:57.051249 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfdll" event={"ID":"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d","Type":"ContainerDied","Data":"f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f"} Dec 06 08:29:58 crc kubenswrapper[4945]: I1206 08:29:58.061868 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfdll" event={"ID":"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d","Type":"ContainerStarted","Data":"dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7"} Dec 06 08:29:58 crc kubenswrapper[4945]: I1206 08:29:58.086669 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kfdll" podStartSLOduration=4.360010846 podStartE2EDuration="7.086652576s" podCreationTimestamp="2025-12-06 08:29:51 +0000 UTC" firstStartedPulling="2025-12-06 08:29:54.998418962 +0000 UTC m=+5808.453280006" lastFinishedPulling="2025-12-06 08:29:57.725060692 +0000 UTC m=+5811.179921736" observedRunningTime="2025-12-06 08:29:58.081295683 +0000 UTC m=+5811.536156727" watchObservedRunningTime="2025-12-06 08:29:58.086652576 +0000 UTC m=+5811.541513620" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.156474 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz"] Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.158837 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.161119 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.161181 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.166803 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz"] Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.269221 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b16f13b-4a09-4bad-b647-655d12bd88c4-secret-volume\") pod \"collect-profiles-29416830-sc8dz\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.269309 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9n88\" (UniqueName: \"kubernetes.io/projected/4b16f13b-4a09-4bad-b647-655d12bd88c4-kube-api-access-f9n88\") pod \"collect-profiles-29416830-sc8dz\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.269415 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b16f13b-4a09-4bad-b647-655d12bd88c4-config-volume\") pod \"collect-profiles-29416830-sc8dz\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.370979 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b16f13b-4a09-4bad-b647-655d12bd88c4-config-volume\") pod \"collect-profiles-29416830-sc8dz\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.371134 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b16f13b-4a09-4bad-b647-655d12bd88c4-secret-volume\") pod \"collect-profiles-29416830-sc8dz\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.371163 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9n88\" (UniqueName: \"kubernetes.io/projected/4b16f13b-4a09-4bad-b647-655d12bd88c4-kube-api-access-f9n88\") pod \"collect-profiles-29416830-sc8dz\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.371963 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b16f13b-4a09-4bad-b647-655d12bd88c4-config-volume\") pod \"collect-profiles-29416830-sc8dz\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.380487 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b16f13b-4a09-4bad-b647-655d12bd88c4-secret-volume\") pod \"collect-profiles-29416830-sc8dz\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.408184 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9n88\" (UniqueName: \"kubernetes.io/projected/4b16f13b-4a09-4bad-b647-655d12bd88c4-kube-api-access-f9n88\") pod \"collect-profiles-29416830-sc8dz\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:00 crc kubenswrapper[4945]: I1206 08:30:00.510396 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:01 crc kubenswrapper[4945]: I1206 08:30:01.001599 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz"] Dec 06 08:30:01 crc kubenswrapper[4945]: I1206 08:30:01.101853 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f1d750f-7e4b-487c-a552-5bf7c183c7ad","Type":"ContainerStarted","Data":"3f5286bbe44faf07a15542dfa3fd343b31914ac27e649df28930226cfb6aa9c5"} Dec 06 08:30:01 crc kubenswrapper[4945]: I1206 08:30:01.104806 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" event={"ID":"4b16f13b-4a09-4bad-b647-655d12bd88c4","Type":"ContainerStarted","Data":"d1f528bcc7b215f4e0ce710fcc7e3bd4284f340b102ae9a5aeda391ddbeafae9"} Dec 06 08:30:01 crc kubenswrapper[4945]: I1206 08:30:01.107525 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c654f23-0692-4d36-918c-3e669d026c44","Type":"ContainerStarted","Data":"9e63c03dd3cba21b28801bcebffea7968c9b29e7fd71533494a793ffa44d707e"} Dec 06 08:30:01 crc kubenswrapper[4945]: I1206 08:30:01.917134 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:30:01 crc kubenswrapper[4945]: I1206 08:30:01.917184 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:30:01 crc kubenswrapper[4945]: I1206 08:30:01.964520 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:30:02 crc kubenswrapper[4945]: I1206 08:30:02.117458 4945 generic.go:334] "Generic (PLEG): container finished" podID="4b16f13b-4a09-4bad-b647-655d12bd88c4" containerID="3a19e0175f87a94f5da4ecb02686c5b6de18e7178d44591094410dbc86ca6bf7" exitCode=0 Dec 06 08:30:02 crc kubenswrapper[4945]: I1206 08:30:02.117586 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" event={"ID":"4b16f13b-4a09-4bad-b647-655d12bd88c4","Type":"ContainerDied","Data":"3a19e0175f87a94f5da4ecb02686c5b6de18e7178d44591094410dbc86ca6bf7"} Dec 06 08:30:02 crc kubenswrapper[4945]: I1206 08:30:02.171828 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:30:02 crc kubenswrapper[4945]: I1206 08:30:02.226156 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kfdll"] Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.125139 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.568026 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.732972 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b16f13b-4a09-4bad-b647-655d12bd88c4-config-volume\") pod \"4b16f13b-4a09-4bad-b647-655d12bd88c4\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.733069 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9n88\" (UniqueName: \"kubernetes.io/projected/4b16f13b-4a09-4bad-b647-655d12bd88c4-kube-api-access-f9n88\") pod \"4b16f13b-4a09-4bad-b647-655d12bd88c4\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.733226 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b16f13b-4a09-4bad-b647-655d12bd88c4-secret-volume\") pod \"4b16f13b-4a09-4bad-b647-655d12bd88c4\" (UID: \"4b16f13b-4a09-4bad-b647-655d12bd88c4\") " Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.735702 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b16f13b-4a09-4bad-b647-655d12bd88c4-config-volume" (OuterVolumeSpecName: "config-volume") pod "4b16f13b-4a09-4bad-b647-655d12bd88c4" (UID: "4b16f13b-4a09-4bad-b647-655d12bd88c4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.740997 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b16f13b-4a09-4bad-b647-655d12bd88c4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4b16f13b-4a09-4bad-b647-655d12bd88c4" (UID: "4b16f13b-4a09-4bad-b647-655d12bd88c4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.745533 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b16f13b-4a09-4bad-b647-655d12bd88c4-kube-api-access-f9n88" (OuterVolumeSpecName: "kube-api-access-f9n88") pod "4b16f13b-4a09-4bad-b647-655d12bd88c4" (UID: "4b16f13b-4a09-4bad-b647-655d12bd88c4"). InnerVolumeSpecName "kube-api-access-f9n88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.843091 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b16f13b-4a09-4bad-b647-655d12bd88c4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.843137 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9n88\" (UniqueName: \"kubernetes.io/projected/4b16f13b-4a09-4bad-b647-655d12bd88c4-kube-api-access-f9n88\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:03 crc kubenswrapper[4945]: I1206 08:30:03.843149 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b16f13b-4a09-4bad-b647-655d12bd88c4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.137496 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" event={"ID":"4b16f13b-4a09-4bad-b647-655d12bd88c4","Type":"ContainerDied","Data":"d1f528bcc7b215f4e0ce710fcc7e3bd4284f340b102ae9a5aeda391ddbeafae9"} Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.137561 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1f528bcc7b215f4e0ce710fcc7e3bd4284f340b102ae9a5aeda391ddbeafae9" Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.137516 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz" Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.137700 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kfdll" podUID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerName="registry-server" containerID="cri-o://dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7" gracePeriod=2 Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.629332 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz"] Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.637693 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416785-znwrz"] Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.675916 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.762407 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-catalog-content\") pod \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.762476 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-utilities\") pod \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.762607 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8w7s\" (UniqueName: \"kubernetes.io/projected/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-kube-api-access-s8w7s\") pod \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\" (UID: \"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d\") " Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.763864 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-utilities" (OuterVolumeSpecName: "utilities") pod "bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" (UID: "bba37ad3-2e7b-4f2e-a557-a35f7e218e8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.768673 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-kube-api-access-s8w7s" (OuterVolumeSpecName: "kube-api-access-s8w7s") pod "bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" (UID: "bba37ad3-2e7b-4f2e-a557-a35f7e218e8d"). InnerVolumeSpecName "kube-api-access-s8w7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.781595 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" (UID: "bba37ad3-2e7b-4f2e-a557-a35f7e218e8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.865080 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8w7s\" (UniqueName: \"kubernetes.io/projected/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-kube-api-access-s8w7s\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.865113 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.865122 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:04 crc kubenswrapper[4945]: I1206 08:30:04.964638 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44485cbb-739b-4d1b-90bd-7a413cff1896" path="/var/lib/kubelet/pods/44485cbb-739b-4d1b-90bd-7a413cff1896/volumes" Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.150441 4945 generic.go:334] "Generic (PLEG): container finished" podID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerID="dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7" exitCode=0 Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.150488 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfdll" event={"ID":"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d","Type":"ContainerDied","Data":"dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7"} Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.150495 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kfdll" Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.150524 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfdll" event={"ID":"bba37ad3-2e7b-4f2e-a557-a35f7e218e8d","Type":"ContainerDied","Data":"2a3a7f7b80ad86ccdbab3dbe71136ebd61c47f01cfb0a599ad4d1ada1360dd8a"} Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.150544 4945 scope.go:117] "RemoveContainer" containerID="dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7" Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.176294 4945 scope.go:117] "RemoveContainer" containerID="f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f" Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.182374 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kfdll"] Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.194288 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kfdll"] Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.210695 4945 scope.go:117] "RemoveContainer" containerID="f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c" Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.241517 4945 scope.go:117] "RemoveContainer" containerID="dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7" Dec 06 08:30:05 crc kubenswrapper[4945]: E1206 08:30:05.242740 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7\": container with ID starting with dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7 not found: ID does not exist" containerID="dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7" Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.242804 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7"} err="failed to get container status \"dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7\": rpc error: code = NotFound desc = could not find container \"dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7\": container with ID starting with dd0aa2dc0823a152321bfb3f067a9022f2b3d334c6f2f00911168c385c5dd9a7 not found: ID does not exist" Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.242843 4945 scope.go:117] "RemoveContainer" containerID="f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f" Dec 06 08:30:05 crc kubenswrapper[4945]: E1206 08:30:05.243607 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f\": container with ID starting with f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f not found: ID does not exist" containerID="f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f" Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.243656 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f"} err="failed to get container status \"f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f\": rpc error: code = NotFound desc = could not find container \"f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f\": container with ID starting with f020140fa4bbd4209b28a65c4ac1d7ced7c6cf8423c9fbdfadad8efcaef2835f not found: ID does not exist" Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.243679 4945 scope.go:117] "RemoveContainer" containerID="f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c" Dec 06 08:30:05 crc kubenswrapper[4945]: E1206 08:30:05.243972 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c\": container with ID starting with f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c not found: ID does not exist" containerID="f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c" Dec 06 08:30:05 crc kubenswrapper[4945]: I1206 08:30:05.244013 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c"} err="failed to get container status \"f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c\": rpc error: code = NotFound desc = could not find container \"f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c\": container with ID starting with f91f19e75113f95fed5803bd7fa3dab9beb9719d43953dcbd5494d752fab490c not found: ID does not exist" Dec 06 08:30:06 crc kubenswrapper[4945]: I1206 08:30:06.968751 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" path="/var/lib/kubelet/pods/bba37ad3-2e7b-4f2e-a557-a35f7e218e8d/volumes" Dec 06 08:30:07 crc kubenswrapper[4945]: I1206 08:30:07.032450 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-8qccw"] Dec 06 08:30:07 crc kubenswrapper[4945]: I1206 08:30:07.044234 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-8qccw"] Dec 06 08:30:07 crc kubenswrapper[4945]: I1206 08:30:07.052520 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b3a2-account-create-update-mn7tt"] Dec 06 08:30:07 crc kubenswrapper[4945]: I1206 08:30:07.060745 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b3a2-account-create-update-mn7tt"] Dec 06 08:30:07 crc kubenswrapper[4945]: I1206 08:30:07.190474 4945 generic.go:334] "Generic (PLEG): container finished" podID="4c654f23-0692-4d36-918c-3e669d026c44" containerID="9e63c03dd3cba21b28801bcebffea7968c9b29e7fd71533494a793ffa44d707e" exitCode=0 Dec 06 08:30:07 crc kubenswrapper[4945]: I1206 08:30:07.190518 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c654f23-0692-4d36-918c-3e669d026c44","Type":"ContainerDied","Data":"9e63c03dd3cba21b28801bcebffea7968c9b29e7fd71533494a793ffa44d707e"} Dec 06 08:30:08 crc kubenswrapper[4945]: I1206 08:30:08.200191 4945 generic.go:334] "Generic (PLEG): container finished" podID="5f1d750f-7e4b-487c-a552-5bf7c183c7ad" containerID="3f5286bbe44faf07a15542dfa3fd343b31914ac27e649df28930226cfb6aa9c5" exitCode=0 Dec 06 08:30:08 crc kubenswrapper[4945]: I1206 08:30:08.200229 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f1d750f-7e4b-487c-a552-5bf7c183c7ad","Type":"ContainerDied","Data":"3f5286bbe44faf07a15542dfa3fd343b31914ac27e649df28930226cfb6aa9c5"} Dec 06 08:30:08 crc kubenswrapper[4945]: I1206 08:30:08.972953 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c788c88-bfb8-4ce3-8316-3ad663f44602" path="/var/lib/kubelet/pods/3c788c88-bfb8-4ce3-8316-3ad663f44602/volumes" Dec 06 08:30:08 crc kubenswrapper[4945]: I1206 08:30:08.974384 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a32bacdd-7760-40d0-ae80-b16857e2c03a" path="/var/lib/kubelet/pods/a32bacdd-7760-40d0-ae80-b16857e2c03a/volumes" Dec 06 08:30:15 crc kubenswrapper[4945]: I1206 08:30:15.271429 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f1d750f-7e4b-487c-a552-5bf7c183c7ad","Type":"ContainerStarted","Data":"85000414026dc1a17900d70b0a1f04634b93eaa6dfdf0cb68a84ae4f4ed9a3c1"} Dec 06 08:30:15 crc kubenswrapper[4945]: I1206 08:30:15.275152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c654f23-0692-4d36-918c-3e669d026c44","Type":"ContainerStarted","Data":"aa15606e2e3c1c6efb41bf0adbc8888c85c05070cc5cf6c5f26613d502eb52f3"} Dec 06 08:30:18 crc kubenswrapper[4945]: I1206 08:30:18.303688 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c654f23-0692-4d36-918c-3e669d026c44","Type":"ContainerStarted","Data":"c7fa7e851e665921c498b65befebdc71af59015924d6ca4b157a94dd77401b94"} Dec 06 08:30:19 crc kubenswrapper[4945]: I1206 08:30:19.314072 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f1d750f-7e4b-487c-a552-5bf7c183c7ad","Type":"ContainerStarted","Data":"388869295a73c956fb1d88e20df0e4cbc7b3cdf4ecc2d6ddba4bb1a6acff0f1e"} Dec 06 08:30:19 crc kubenswrapper[4945]: I1206 08:30:19.315787 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 06 08:30:19 crc kubenswrapper[4945]: I1206 08:30:19.321669 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 06 08:30:19 crc kubenswrapper[4945]: I1206 08:30:19.373130 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.453348614 podStartE2EDuration="26.373110381s" podCreationTimestamp="2025-12-06 08:29:53 +0000 UTC" firstStartedPulling="2025-12-06 08:29:54.912245756 +0000 UTC m=+5808.367106800" lastFinishedPulling="2025-12-06 08:30:14.832007523 +0000 UTC m=+5828.286868567" observedRunningTime="2025-12-06 08:30:19.344063194 +0000 UTC m=+5832.798924238" watchObservedRunningTime="2025-12-06 08:30:19.373110381 +0000 UTC m=+5832.827971425" Dec 06 08:30:21 crc kubenswrapper[4945]: I1206 08:30:21.333549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c654f23-0692-4d36-918c-3e669d026c44","Type":"ContainerStarted","Data":"a31466b020653584f91c956689a0d12ca6c88f64ee73683681738500626865ad"} Dec 06 08:30:21 crc kubenswrapper[4945]: I1206 08:30:21.373082 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.253998574 podStartE2EDuration="28.373064059s" podCreationTimestamp="2025-12-06 08:29:53 +0000 UTC" firstStartedPulling="2025-12-06 08:29:55.268397905 +0000 UTC m=+5808.723258949" lastFinishedPulling="2025-12-06 08:30:20.38746339 +0000 UTC m=+5833.842324434" observedRunningTime="2025-12-06 08:30:21.370811888 +0000 UTC m=+5834.825672942" watchObservedRunningTime="2025-12-06 08:30:21.373064059 +0000 UTC m=+5834.827925113" Dec 06 08:30:24 crc kubenswrapper[4945]: I1206 08:30:24.632916 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 06 08:30:24 crc kubenswrapper[4945]: I1206 08:30:24.633212 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 06 08:30:24 crc kubenswrapper[4945]: I1206 08:30:24.636957 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 06 08:30:25 crc kubenswrapper[4945]: I1206 08:30:25.386576 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 06 08:30:26 crc kubenswrapper[4945]: I1206 08:30:26.313470 4945 scope.go:117] "RemoveContainer" containerID="1274ff27990c32dfd585c75d16dfab0e5d2a1d36f12b57aa1883e053c33dfc54" Dec 06 08:30:26 crc kubenswrapper[4945]: I1206 08:30:26.360253 4945 scope.go:117] "RemoveContainer" containerID="d980bc39cee707bd1d99b68e271d88cf072743d74d782f674b630186e1297e0c" Dec 06 08:30:26 crc kubenswrapper[4945]: I1206 08:30:26.391109 4945 scope.go:117] "RemoveContainer" containerID="5b8dcd82f9b3b9ca89221d42f990e087e941cef295251973dd53c05c2aec8aa7" Dec 06 08:30:26 crc kubenswrapper[4945]: I1206 08:30:26.436216 4945 scope.go:117] "RemoveContainer" containerID="ac7364973e8403f04634ad22bb330c62c86967a21c8c84bef280f0792f020922" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.086869 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:30:27 crc kubenswrapper[4945]: E1206 08:30:27.090630 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerName="extract-utilities" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.090661 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerName="extract-utilities" Dec 06 08:30:27 crc kubenswrapper[4945]: E1206 08:30:27.090678 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerName="registry-server" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.090685 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerName="registry-server" Dec 06 08:30:27 crc kubenswrapper[4945]: E1206 08:30:27.090695 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerName="extract-content" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.090702 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerName="extract-content" Dec 06 08:30:27 crc kubenswrapper[4945]: E1206 08:30:27.090723 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b16f13b-4a09-4bad-b647-655d12bd88c4" containerName="collect-profiles" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.090729 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b16f13b-4a09-4bad-b647-655d12bd88c4" containerName="collect-profiles" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.090975 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bba37ad3-2e7b-4f2e-a557-a35f7e218e8d" containerName="registry-server" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.090997 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b16f13b-4a09-4bad-b647-655d12bd88c4" containerName="collect-profiles" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.092910 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.095540 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.095646 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.110994 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.262165 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4drf\" (UniqueName: \"kubernetes.io/projected/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-kube-api-access-c4drf\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.262473 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-config-data\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.262642 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-run-httpd\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.262781 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-log-httpd\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.265335 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.265426 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-scripts\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.265509 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.366922 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-log-httpd\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.368923 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.369139 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-scripts\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.369331 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.367367 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-log-httpd\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.369534 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4drf\" (UniqueName: \"kubernetes.io/projected/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-kube-api-access-c4drf\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.369689 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-config-data\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.369843 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-run-httpd\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.375827 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-run-httpd\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.375906 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.376235 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-scripts\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.377386 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.380521 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-config-data\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.385603 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4drf\" (UniqueName: \"kubernetes.io/projected/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-kube-api-access-c4drf\") pod \"ceilometer-0\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.414566 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 08:30:27 crc kubenswrapper[4945]: I1206 08:30:27.987058 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:30:27 crc kubenswrapper[4945]: W1206 08:30:27.987980 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbe0274d_42a7_4730_ae93_3fbb07e6e8c8.slice/crio-1911f88efb772c003b9e23df13c5720bc5640ad8266d8ed87e9407636ffd69b4 WatchSource:0}: Error finding container 1911f88efb772c003b9e23df13c5720bc5640ad8266d8ed87e9407636ffd69b4: Status 404 returned error can't find the container with id 1911f88efb772c003b9e23df13c5720bc5640ad8266d8ed87e9407636ffd69b4 Dec 06 08:30:28 crc kubenswrapper[4945]: I1206 08:30:28.417709 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8","Type":"ContainerStarted","Data":"1911f88efb772c003b9e23df13c5720bc5640ad8266d8ed87e9407636ffd69b4"} Dec 06 08:30:32 crc kubenswrapper[4945]: I1206 08:30:32.051088 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-nnq2k"] Dec 06 08:30:32 crc kubenswrapper[4945]: I1206 08:30:32.065001 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-nnq2k"] Dec 06 08:30:32 crc kubenswrapper[4945]: I1206 08:30:32.460144 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8","Type":"ContainerStarted","Data":"4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57"} Dec 06 08:30:32 crc kubenswrapper[4945]: I1206 08:30:32.964731 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eb4c4a7-7e65-444c-b65b-206632d5037b" path="/var/lib/kubelet/pods/7eb4c4a7-7e65-444c-b65b-206632d5037b/volumes" Dec 06 08:30:33 crc kubenswrapper[4945]: I1206 08:30:33.469354 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8","Type":"ContainerStarted","Data":"2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a"} Dec 06 08:30:34 crc kubenswrapper[4945]: I1206 08:30:34.481645 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8","Type":"ContainerStarted","Data":"9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f"} Dec 06 08:30:36 crc kubenswrapper[4945]: I1206 08:30:36.499290 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8","Type":"ContainerStarted","Data":"38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851"} Dec 06 08:30:36 crc kubenswrapper[4945]: I1206 08:30:36.500860 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 08:30:39 crc kubenswrapper[4945]: I1206 08:30:39.870510 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.4306945429999995 podStartE2EDuration="12.870486652s" podCreationTimestamp="2025-12-06 08:30:27 +0000 UTC" firstStartedPulling="2025-12-06 08:30:27.990270861 +0000 UTC m=+5841.445131905" lastFinishedPulling="2025-12-06 08:30:35.43006297 +0000 UTC m=+5848.884924014" observedRunningTime="2025-12-06 08:30:36.522745404 +0000 UTC m=+5849.977606458" watchObservedRunningTime="2025-12-06 08:30:39.870486652 +0000 UTC m=+5853.325347696" Dec 06 08:30:39 crc kubenswrapper[4945]: I1206 08:30:39.878866 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-njgmq"] Dec 06 08:30:39 crc kubenswrapper[4945]: I1206 08:30:39.880266 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-njgmq" Dec 06 08:30:39 crc kubenswrapper[4945]: I1206 08:30:39.892155 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-njgmq"] Dec 06 08:30:39 crc kubenswrapper[4945]: I1206 08:30:39.972391 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-8532-account-create-update-ch9cf"] Dec 06 08:30:39 crc kubenswrapper[4945]: I1206 08:30:39.973730 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8532-account-create-update-ch9cf" Dec 06 08:30:39 crc kubenswrapper[4945]: I1206 08:30:39.975582 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 06 08:30:39 crc kubenswrapper[4945]: I1206 08:30:39.991323 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-8532-account-create-update-ch9cf"] Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.043780 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9753090-1060-47a1-b7fd-86c494da37e7-operator-scripts\") pod \"aodh-db-create-njgmq\" (UID: \"c9753090-1060-47a1-b7fd-86c494da37e7\") " pod="openstack/aodh-db-create-njgmq" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.043882 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgcjn\" (UniqueName: \"kubernetes.io/projected/c9753090-1060-47a1-b7fd-86c494da37e7-kube-api-access-hgcjn\") pod \"aodh-db-create-njgmq\" (UID: \"c9753090-1060-47a1-b7fd-86c494da37e7\") " pod="openstack/aodh-db-create-njgmq" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.145505 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c5m4\" (UniqueName: \"kubernetes.io/projected/14bc7530-6182-468d-9d49-ef02db905104-kube-api-access-9c5m4\") pod \"aodh-8532-account-create-update-ch9cf\" (UID: \"14bc7530-6182-468d-9d49-ef02db905104\") " pod="openstack/aodh-8532-account-create-update-ch9cf" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.145785 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9753090-1060-47a1-b7fd-86c494da37e7-operator-scripts\") pod \"aodh-db-create-njgmq\" (UID: \"c9753090-1060-47a1-b7fd-86c494da37e7\") " pod="openstack/aodh-db-create-njgmq" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.146019 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgcjn\" (UniqueName: \"kubernetes.io/projected/c9753090-1060-47a1-b7fd-86c494da37e7-kube-api-access-hgcjn\") pod \"aodh-db-create-njgmq\" (UID: \"c9753090-1060-47a1-b7fd-86c494da37e7\") " pod="openstack/aodh-db-create-njgmq" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.146063 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14bc7530-6182-468d-9d49-ef02db905104-operator-scripts\") pod \"aodh-8532-account-create-update-ch9cf\" (UID: \"14bc7530-6182-468d-9d49-ef02db905104\") " pod="openstack/aodh-8532-account-create-update-ch9cf" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.146507 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9753090-1060-47a1-b7fd-86c494da37e7-operator-scripts\") pod \"aodh-db-create-njgmq\" (UID: \"c9753090-1060-47a1-b7fd-86c494da37e7\") " pod="openstack/aodh-db-create-njgmq" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.184517 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgcjn\" (UniqueName: \"kubernetes.io/projected/c9753090-1060-47a1-b7fd-86c494da37e7-kube-api-access-hgcjn\") pod \"aodh-db-create-njgmq\" (UID: \"c9753090-1060-47a1-b7fd-86c494da37e7\") " pod="openstack/aodh-db-create-njgmq" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.221566 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-njgmq" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.248302 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14bc7530-6182-468d-9d49-ef02db905104-operator-scripts\") pod \"aodh-8532-account-create-update-ch9cf\" (UID: \"14bc7530-6182-468d-9d49-ef02db905104\") " pod="openstack/aodh-8532-account-create-update-ch9cf" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.248556 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c5m4\" (UniqueName: \"kubernetes.io/projected/14bc7530-6182-468d-9d49-ef02db905104-kube-api-access-9c5m4\") pod \"aodh-8532-account-create-update-ch9cf\" (UID: \"14bc7530-6182-468d-9d49-ef02db905104\") " pod="openstack/aodh-8532-account-create-update-ch9cf" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.249174 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14bc7530-6182-468d-9d49-ef02db905104-operator-scripts\") pod \"aodh-8532-account-create-update-ch9cf\" (UID: \"14bc7530-6182-468d-9d49-ef02db905104\") " pod="openstack/aodh-8532-account-create-update-ch9cf" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.274336 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c5m4\" (UniqueName: \"kubernetes.io/projected/14bc7530-6182-468d-9d49-ef02db905104-kube-api-access-9c5m4\") pod \"aodh-8532-account-create-update-ch9cf\" (UID: \"14bc7530-6182-468d-9d49-ef02db905104\") " pod="openstack/aodh-8532-account-create-update-ch9cf" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.290853 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8532-account-create-update-ch9cf" Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.786340 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-njgmq"] Dec 06 08:30:40 crc kubenswrapper[4945]: W1206 08:30:40.790990 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9753090_1060_47a1_b7fd_86c494da37e7.slice/crio-66af8e01adf983b202915e9b9b851a7f670b82cd75691eeaf9a8a4f702e3dd6d WatchSource:0}: Error finding container 66af8e01adf983b202915e9b9b851a7f670b82cd75691eeaf9a8a4f702e3dd6d: Status 404 returned error can't find the container with id 66af8e01adf983b202915e9b9b851a7f670b82cd75691eeaf9a8a4f702e3dd6d Dec 06 08:30:40 crc kubenswrapper[4945]: I1206 08:30:40.875900 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-8532-account-create-update-ch9cf"] Dec 06 08:30:41 crc kubenswrapper[4945]: I1206 08:30:41.544195 4945 generic.go:334] "Generic (PLEG): container finished" podID="14bc7530-6182-468d-9d49-ef02db905104" containerID="19e5ed44e570fd355b1c520049c965a375defab39bc1f58986a6fa70cd598df2" exitCode=0 Dec 06 08:30:41 crc kubenswrapper[4945]: I1206 08:30:41.544249 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-8532-account-create-update-ch9cf" event={"ID":"14bc7530-6182-468d-9d49-ef02db905104","Type":"ContainerDied","Data":"19e5ed44e570fd355b1c520049c965a375defab39bc1f58986a6fa70cd598df2"} Dec 06 08:30:41 crc kubenswrapper[4945]: I1206 08:30:41.544319 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-8532-account-create-update-ch9cf" event={"ID":"14bc7530-6182-468d-9d49-ef02db905104","Type":"ContainerStarted","Data":"06295c1f64b05167abb8ce54456a3c06947f0f2e299597f4d0599d0340353c39"} Dec 06 08:30:41 crc kubenswrapper[4945]: I1206 08:30:41.545828 4945 generic.go:334] "Generic (PLEG): container finished" podID="c9753090-1060-47a1-b7fd-86c494da37e7" containerID="493728116c44da3769a9dd432e913aba3b3693edea6960e644f640add3eaa5b2" exitCode=0 Dec 06 08:30:41 crc kubenswrapper[4945]: I1206 08:30:41.545863 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-njgmq" event={"ID":"c9753090-1060-47a1-b7fd-86c494da37e7","Type":"ContainerDied","Data":"493728116c44da3769a9dd432e913aba3b3693edea6960e644f640add3eaa5b2"} Dec 06 08:30:41 crc kubenswrapper[4945]: I1206 08:30:41.545878 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-njgmq" event={"ID":"c9753090-1060-47a1-b7fd-86c494da37e7","Type":"ContainerStarted","Data":"66af8e01adf983b202915e9b9b851a7f670b82cd75691eeaf9a8a4f702e3dd6d"} Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.057501 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-njgmq" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.066661 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8532-account-create-update-ch9cf" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.208368 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14bc7530-6182-468d-9d49-ef02db905104-operator-scripts\") pod \"14bc7530-6182-468d-9d49-ef02db905104\" (UID: \"14bc7530-6182-468d-9d49-ef02db905104\") " Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.208585 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgcjn\" (UniqueName: \"kubernetes.io/projected/c9753090-1060-47a1-b7fd-86c494da37e7-kube-api-access-hgcjn\") pod \"c9753090-1060-47a1-b7fd-86c494da37e7\" (UID: \"c9753090-1060-47a1-b7fd-86c494da37e7\") " Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.208667 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9753090-1060-47a1-b7fd-86c494da37e7-operator-scripts\") pod \"c9753090-1060-47a1-b7fd-86c494da37e7\" (UID: \"c9753090-1060-47a1-b7fd-86c494da37e7\") " Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.208725 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c5m4\" (UniqueName: \"kubernetes.io/projected/14bc7530-6182-468d-9d49-ef02db905104-kube-api-access-9c5m4\") pod \"14bc7530-6182-468d-9d49-ef02db905104\" (UID: \"14bc7530-6182-468d-9d49-ef02db905104\") " Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.209260 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9753090-1060-47a1-b7fd-86c494da37e7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c9753090-1060-47a1-b7fd-86c494da37e7" (UID: "c9753090-1060-47a1-b7fd-86c494da37e7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.209431 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bc7530-6182-468d-9d49-ef02db905104-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "14bc7530-6182-468d-9d49-ef02db905104" (UID: "14bc7530-6182-468d-9d49-ef02db905104"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.214123 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9753090-1060-47a1-b7fd-86c494da37e7-kube-api-access-hgcjn" (OuterVolumeSpecName: "kube-api-access-hgcjn") pod "c9753090-1060-47a1-b7fd-86c494da37e7" (UID: "c9753090-1060-47a1-b7fd-86c494da37e7"). InnerVolumeSpecName "kube-api-access-hgcjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.214215 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14bc7530-6182-468d-9d49-ef02db905104-kube-api-access-9c5m4" (OuterVolumeSpecName: "kube-api-access-9c5m4") pod "14bc7530-6182-468d-9d49-ef02db905104" (UID: "14bc7530-6182-468d-9d49-ef02db905104"). InnerVolumeSpecName "kube-api-access-9c5m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.311067 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14bc7530-6182-468d-9d49-ef02db905104-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.311116 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgcjn\" (UniqueName: \"kubernetes.io/projected/c9753090-1060-47a1-b7fd-86c494da37e7-kube-api-access-hgcjn\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.311133 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9753090-1060-47a1-b7fd-86c494da37e7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.311145 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c5m4\" (UniqueName: \"kubernetes.io/projected/14bc7530-6182-468d-9d49-ef02db905104-kube-api-access-9c5m4\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.577135 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-8532-account-create-update-ch9cf" event={"ID":"14bc7530-6182-468d-9d49-ef02db905104","Type":"ContainerDied","Data":"06295c1f64b05167abb8ce54456a3c06947f0f2e299597f4d0599d0340353c39"} Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.577174 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8532-account-create-update-ch9cf" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.577197 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06295c1f64b05167abb8ce54456a3c06947f0f2e299597f4d0599d0340353c39" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.578692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-njgmq" event={"ID":"c9753090-1060-47a1-b7fd-86c494da37e7","Type":"ContainerDied","Data":"66af8e01adf983b202915e9b9b851a7f670b82cd75691eeaf9a8a4f702e3dd6d"} Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.578729 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66af8e01adf983b202915e9b9b851a7f670b82cd75691eeaf9a8a4f702e3dd6d" Dec 06 08:30:43 crc kubenswrapper[4945]: I1206 08:30:43.578773 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-njgmq" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.349060 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-9fzp9"] Dec 06 08:30:45 crc kubenswrapper[4945]: E1206 08:30:45.349872 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14bc7530-6182-468d-9d49-ef02db905104" containerName="mariadb-account-create-update" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.349889 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="14bc7530-6182-468d-9d49-ef02db905104" containerName="mariadb-account-create-update" Dec 06 08:30:45 crc kubenswrapper[4945]: E1206 08:30:45.349903 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9753090-1060-47a1-b7fd-86c494da37e7" containerName="mariadb-database-create" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.349909 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9753090-1060-47a1-b7fd-86c494da37e7" containerName="mariadb-database-create" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.350113 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9753090-1060-47a1-b7fd-86c494da37e7" containerName="mariadb-database-create" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.350145 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="14bc7530-6182-468d-9d49-ef02db905104" containerName="mariadb-account-create-update" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.350860 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.353428 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.353701 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.353782 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.353958 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-99p7g" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.363805 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-9fzp9"] Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.463664 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-scripts\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.463711 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n28vg\" (UniqueName: \"kubernetes.io/projected/978dc61d-451e-49e8-9cc4-84f49906caae-kube-api-access-n28vg\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.463761 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-config-data\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.463836 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-combined-ca-bundle\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.564954 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-combined-ca-bundle\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.565051 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-scripts\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.565085 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n28vg\" (UniqueName: \"kubernetes.io/projected/978dc61d-451e-49e8-9cc4-84f49906caae-kube-api-access-n28vg\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.565139 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-config-data\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.570776 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-scripts\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.571209 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-config-data\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.573459 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-combined-ca-bundle\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.591016 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n28vg\" (UniqueName: \"kubernetes.io/projected/978dc61d-451e-49e8-9cc4-84f49906caae-kube-api-access-n28vg\") pod \"aodh-db-sync-9fzp9\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:45 crc kubenswrapper[4945]: I1206 08:30:45.681997 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:46 crc kubenswrapper[4945]: W1206 08:30:46.171685 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod978dc61d_451e_49e8_9cc4_84f49906caae.slice/crio-4442a5291070c71e6e9481146ba04ddf64047d795ba850ab2cb6d548ec35f546 WatchSource:0}: Error finding container 4442a5291070c71e6e9481146ba04ddf64047d795ba850ab2cb6d548ec35f546: Status 404 returned error can't find the container with id 4442a5291070c71e6e9481146ba04ddf64047d795ba850ab2cb6d548ec35f546 Dec 06 08:30:46 crc kubenswrapper[4945]: I1206 08:30:46.172367 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-9fzp9"] Dec 06 08:30:46 crc kubenswrapper[4945]: I1206 08:30:46.607810 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-9fzp9" event={"ID":"978dc61d-451e-49e8-9cc4-84f49906caae","Type":"ContainerStarted","Data":"4442a5291070c71e6e9481146ba04ddf64047d795ba850ab2cb6d548ec35f546"} Dec 06 08:30:53 crc kubenswrapper[4945]: I1206 08:30:53.677980 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-9fzp9" event={"ID":"978dc61d-451e-49e8-9cc4-84f49906caae","Type":"ContainerStarted","Data":"37f9d2d64fb604dceb981e0071c548bfad58b7354e15b4ee927046c9521f86f6"} Dec 06 08:30:53 crc kubenswrapper[4945]: I1206 08:30:53.697214 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-9fzp9" podStartSLOduration=2.111442562 podStartE2EDuration="8.697199102s" podCreationTimestamp="2025-12-06 08:30:45 +0000 UTC" firstStartedPulling="2025-12-06 08:30:46.174834564 +0000 UTC m=+5859.629695598" lastFinishedPulling="2025-12-06 08:30:52.760591094 +0000 UTC m=+5866.215452138" observedRunningTime="2025-12-06 08:30:53.696125063 +0000 UTC m=+5867.150986107" watchObservedRunningTime="2025-12-06 08:30:53.697199102 +0000 UTC m=+5867.152060146" Dec 06 08:30:55 crc kubenswrapper[4945]: I1206 08:30:55.700568 4945 generic.go:334] "Generic (PLEG): container finished" podID="978dc61d-451e-49e8-9cc4-84f49906caae" containerID="37f9d2d64fb604dceb981e0071c548bfad58b7354e15b4ee927046c9521f86f6" exitCode=0 Dec 06 08:30:55 crc kubenswrapper[4945]: I1206 08:30:55.700683 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-9fzp9" event={"ID":"978dc61d-451e-49e8-9cc4-84f49906caae","Type":"ContainerDied","Data":"37f9d2d64fb604dceb981e0071c548bfad58b7354e15b4ee927046c9521f86f6"} Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.079959 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.206869 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n28vg\" (UniqueName: \"kubernetes.io/projected/978dc61d-451e-49e8-9cc4-84f49906caae-kube-api-access-n28vg\") pod \"978dc61d-451e-49e8-9cc4-84f49906caae\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.207027 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-scripts\") pod \"978dc61d-451e-49e8-9cc4-84f49906caae\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.207076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-config-data\") pod \"978dc61d-451e-49e8-9cc4-84f49906caae\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.207138 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-combined-ca-bundle\") pod \"978dc61d-451e-49e8-9cc4-84f49906caae\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.212638 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/978dc61d-451e-49e8-9cc4-84f49906caae-kube-api-access-n28vg" (OuterVolumeSpecName: "kube-api-access-n28vg") pod "978dc61d-451e-49e8-9cc4-84f49906caae" (UID: "978dc61d-451e-49e8-9cc4-84f49906caae"). InnerVolumeSpecName "kube-api-access-n28vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.212689 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-scripts" (OuterVolumeSpecName: "scripts") pod "978dc61d-451e-49e8-9cc4-84f49906caae" (UID: "978dc61d-451e-49e8-9cc4-84f49906caae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:30:57 crc kubenswrapper[4945]: E1206 08:30:57.237536 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-config-data podName:978dc61d-451e-49e8-9cc4-84f49906caae nodeName:}" failed. No retries permitted until 2025-12-06 08:30:57.737505273 +0000 UTC m=+5871.192366327 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-config-data") pod "978dc61d-451e-49e8-9cc4-84f49906caae" (UID: "978dc61d-451e-49e8-9cc4-84f49906caae") : error deleting /var/lib/kubelet/pods/978dc61d-451e-49e8-9cc4-84f49906caae/volume-subpaths: remove /var/lib/kubelet/pods/978dc61d-451e-49e8-9cc4-84f49906caae/volume-subpaths: no such file or directory Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.240407 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "978dc61d-451e-49e8-9cc4-84f49906caae" (UID: "978dc61d-451e-49e8-9cc4-84f49906caae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.309681 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.309712 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.309801 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n28vg\" (UniqueName: \"kubernetes.io/projected/978dc61d-451e-49e8-9cc4-84f49906caae-kube-api-access-n28vg\") on node \"crc\" DevicePath \"\"" Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.424149 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.729264 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-9fzp9" event={"ID":"978dc61d-451e-49e8-9cc4-84f49906caae","Type":"ContainerDied","Data":"4442a5291070c71e6e9481146ba04ddf64047d795ba850ab2cb6d548ec35f546"} Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.729316 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4442a5291070c71e6e9481146ba04ddf64047d795ba850ab2cb6d548ec35f546" Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.729324 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-9fzp9" Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.818838 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-config-data\") pod \"978dc61d-451e-49e8-9cc4-84f49906caae\" (UID: \"978dc61d-451e-49e8-9cc4-84f49906caae\") " Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.823516 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-config-data" (OuterVolumeSpecName: "config-data") pod "978dc61d-451e-49e8-9cc4-84f49906caae" (UID: "978dc61d-451e-49e8-9cc4-84f49906caae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:30:57 crc kubenswrapper[4945]: I1206 08:30:57.921084 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978dc61d-451e-49e8-9cc4-84f49906caae-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.483224 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 06 08:31:00 crc kubenswrapper[4945]: E1206 08:31:00.484221 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="978dc61d-451e-49e8-9cc4-84f49906caae" containerName="aodh-db-sync" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.484242 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="978dc61d-451e-49e8-9cc4-84f49906caae" containerName="aodh-db-sync" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.485010 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="978dc61d-451e-49e8-9cc4-84f49906caae" containerName="aodh-db-sync" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.487471 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.492601 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.492806 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.492907 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-99p7g" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.537825 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.577721 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-config-data\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.577787 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8zc4\" (UniqueName: \"kubernetes.io/projected/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-kube-api-access-z8zc4\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.577980 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-scripts\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.578081 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-combined-ca-bundle\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.679825 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-combined-ca-bundle\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.680786 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-config-data\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.680837 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8zc4\" (UniqueName: \"kubernetes.io/projected/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-kube-api-access-z8zc4\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.680915 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-scripts\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.689371 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-combined-ca-bundle\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.691027 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-scripts\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.700665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8zc4\" (UniqueName: \"kubernetes.io/projected/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-kube-api-access-z8zc4\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.702113 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07-config-data\") pod \"aodh-0\" (UID: \"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07\") " pod="openstack/aodh-0" Dec 06 08:31:00 crc kubenswrapper[4945]: I1206 08:31:00.821331 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 08:31:01 crc kubenswrapper[4945]: I1206 08:31:01.323509 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 06 08:31:01 crc kubenswrapper[4945]: I1206 08:31:01.771177 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07","Type":"ContainerStarted","Data":"f454c5a5d73173e83baff7256ac5e88749e51cc3e7e7603999a604c56e46e121"} Dec 06 08:31:01 crc kubenswrapper[4945]: I1206 08:31:01.772267 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07","Type":"ContainerStarted","Data":"f49d85ead8ef6dea1267d0381b39587f920b2212d0aee8e0c072b120b505eec0"} Dec 06 08:31:02 crc kubenswrapper[4945]: I1206 08:31:02.967854 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:02 crc kubenswrapper[4945]: I1206 08:31:02.968427 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="ceilometer-central-agent" containerID="cri-o://4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57" gracePeriod=30 Dec 06 08:31:02 crc kubenswrapper[4945]: I1206 08:31:02.968501 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="proxy-httpd" containerID="cri-o://38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851" gracePeriod=30 Dec 06 08:31:02 crc kubenswrapper[4945]: I1206 08:31:02.968548 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="ceilometer-notification-agent" containerID="cri-o://2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a" gracePeriod=30 Dec 06 08:31:02 crc kubenswrapper[4945]: I1206 08:31:02.968533 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="sg-core" containerID="cri-o://9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f" gracePeriod=30 Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.038126 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-26dd-account-create-update-h4ps4"] Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.049426 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-vr5x5"] Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.063920 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-vr5x5"] Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.077038 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-26dd-account-create-update-h4ps4"] Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.792928 4945 generic.go:334] "Generic (PLEG): container finished" podID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerID="38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851" exitCode=0 Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.793327 4945 generic.go:334] "Generic (PLEG): container finished" podID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerID="9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f" exitCode=2 Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.793343 4945 generic.go:334] "Generic (PLEG): container finished" podID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerID="4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57" exitCode=0 Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.793006 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8","Type":"ContainerDied","Data":"38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851"} Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.793383 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8","Type":"ContainerDied","Data":"9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f"} Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.793397 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8","Type":"ContainerDied","Data":"4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57"} Dec 06 08:31:03 crc kubenswrapper[4945]: I1206 08:31:03.795705 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07","Type":"ContainerStarted","Data":"a122ace5f164c2c36505ece2edb1e67b5d1104c1cad893108d0201840a716552"} Dec 06 08:31:04 crc kubenswrapper[4945]: I1206 08:31:04.808342 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07","Type":"ContainerStarted","Data":"40591aaa2ab88e86a24a1c1accb55c3e4b85d8e0afbca341cd897e24243fec56"} Dec 06 08:31:04 crc kubenswrapper[4945]: I1206 08:31:04.968269 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="909d1589-58d9-412e-9b0e-15171d896b7b" path="/var/lib/kubelet/pods/909d1589-58d9-412e-9b0e-15171d896b7b/volumes" Dec 06 08:31:04 crc kubenswrapper[4945]: I1206 08:31:04.969427 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1a1ef05-412d-41bb-be03-a2f649984250" path="/var/lib/kubelet/pods/c1a1ef05-412d-41bb-be03-a2f649984250/volumes" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.643198 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.784804 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-log-httpd\") pod \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.785072 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4drf\" (UniqueName: \"kubernetes.io/projected/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-kube-api-access-c4drf\") pod \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.785112 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-sg-core-conf-yaml\") pod \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.785159 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-combined-ca-bundle\") pod \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.785244 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" (UID: "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.785318 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-scripts\") pod \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.785369 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-run-httpd\") pod \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.785401 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-config-data\") pod \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\" (UID: \"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8\") " Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.786168 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.789885 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" (UID: "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.791534 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-kube-api-access-c4drf" (OuterVolumeSpecName: "kube-api-access-c4drf") pod "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" (UID: "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8"). InnerVolumeSpecName "kube-api-access-c4drf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.793119 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-scripts" (OuterVolumeSpecName: "scripts") pod "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" (UID: "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.814985 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" (UID: "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.828159 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07","Type":"ContainerStarted","Data":"cb47a273d03c9c0aa99b1814d1cfd75b25408860955b2a47408af00d68ee4ff7"} Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.835919 4945 generic.go:334] "Generic (PLEG): container finished" podID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerID="2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a" exitCode=0 Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.836020 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.836042 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8","Type":"ContainerDied","Data":"2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a"} Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.836954 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe0274d-42a7-4730-ae93-3fbb07e6e8c8","Type":"ContainerDied","Data":"1911f88efb772c003b9e23df13c5720bc5640ad8266d8ed87e9407636ffd69b4"} Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.836982 4945 scope.go:117] "RemoveContainer" containerID="38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.872337 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.946966315 podStartE2EDuration="5.872312736s" podCreationTimestamp="2025-12-06 08:31:00 +0000 UTC" firstStartedPulling="2025-12-06 08:31:01.330658165 +0000 UTC m=+5874.785519209" lastFinishedPulling="2025-12-06 08:31:05.256004576 +0000 UTC m=+5878.710865630" observedRunningTime="2025-12-06 08:31:05.850168113 +0000 UTC m=+5879.305029157" watchObservedRunningTime="2025-12-06 08:31:05.872312736 +0000 UTC m=+5879.327173790" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.873270 4945 scope.go:117] "RemoveContainer" containerID="9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.898649 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.898684 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.898696 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4drf\" (UniqueName: \"kubernetes.io/projected/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-kube-api-access-c4drf\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.898706 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.915862 4945 scope.go:117] "RemoveContainer" containerID="2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.923118 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" (UID: "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.960806 4945 scope.go:117] "RemoveContainer" containerID="4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57" Dec 06 08:31:05 crc kubenswrapper[4945]: I1206 08:31:05.998246 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-config-data" (OuterVolumeSpecName: "config-data") pod "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" (UID: "bbe0274d-42a7-4730-ae93-3fbb07e6e8c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.001581 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.001739 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.002193 4945 scope.go:117] "RemoveContainer" containerID="38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851" Dec 06 08:31:06 crc kubenswrapper[4945]: E1206 08:31:06.002726 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851\": container with ID starting with 38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851 not found: ID does not exist" containerID="38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.002840 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851"} err="failed to get container status \"38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851\": rpc error: code = NotFound desc = could not find container \"38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851\": container with ID starting with 38ee65671241b8c7e939a1eaf031c4470814b306298b471b55aff32c67b89851 not found: ID does not exist" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.002922 4945 scope.go:117] "RemoveContainer" containerID="9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f" Dec 06 08:31:06 crc kubenswrapper[4945]: E1206 08:31:06.003425 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f\": container with ID starting with 9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f not found: ID does not exist" containerID="9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.003463 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f"} err="failed to get container status \"9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f\": rpc error: code = NotFound desc = could not find container \"9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f\": container with ID starting with 9fc4aca8bd8996052ed874878a4b006729db642de6c29e26cc530fe52bd8a53f not found: ID does not exist" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.003485 4945 scope.go:117] "RemoveContainer" containerID="2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a" Dec 06 08:31:06 crc kubenswrapper[4945]: E1206 08:31:06.003711 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a\": container with ID starting with 2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a not found: ID does not exist" containerID="2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.003824 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a"} err="failed to get container status \"2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a\": rpc error: code = NotFound desc = could not find container \"2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a\": container with ID starting with 2466904d5ebacaa0ed8c7d65f97a90fcf5638becb35e921c1c1f616cb0ae918a not found: ID does not exist" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.003895 4945 scope.go:117] "RemoveContainer" containerID="4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57" Dec 06 08:31:06 crc kubenswrapper[4945]: E1206 08:31:06.004162 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57\": container with ID starting with 4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57 not found: ID does not exist" containerID="4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.004190 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57"} err="failed to get container status \"4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57\": rpc error: code = NotFound desc = could not find container \"4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57\": container with ID starting with 4a7ddc140f3ad32259f7ffd5134fd9d640560e4f9205fcc110cb66bd99144d57 not found: ID does not exist" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.184804 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.203888 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.216205 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:06 crc kubenswrapper[4945]: E1206 08:31:06.216804 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="ceilometer-notification-agent" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.216827 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="ceilometer-notification-agent" Dec 06 08:31:06 crc kubenswrapper[4945]: E1206 08:31:06.216855 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="ceilometer-central-agent" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.216865 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="ceilometer-central-agent" Dec 06 08:31:06 crc kubenswrapper[4945]: E1206 08:31:06.216897 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="sg-core" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.216907 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="sg-core" Dec 06 08:31:06 crc kubenswrapper[4945]: E1206 08:31:06.216925 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="proxy-httpd" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.216934 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="proxy-httpd" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.217247 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="sg-core" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.217309 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="ceilometer-notification-agent" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.217328 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="ceilometer-central-agent" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.217340 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" containerName="proxy-httpd" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.219904 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.229001 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.242662 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.243083 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.309619 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-config-data\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.309687 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c4lj\" (UniqueName: \"kubernetes.io/projected/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-kube-api-access-9c4lj\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.309871 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-log-httpd\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.309917 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.309964 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-scripts\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.310051 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.310106 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-run-httpd\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.411811 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.412193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-run-httpd\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.412258 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-config-data\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.412304 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c4lj\" (UniqueName: \"kubernetes.io/projected/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-kube-api-access-9c4lj\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.412458 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-log-httpd\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.412492 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.412518 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-scripts\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.412732 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-run-httpd\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.413028 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-log-httpd\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.417788 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-config-data\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.418656 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.419572 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.427738 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-scripts\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.438270 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c4lj\" (UniqueName: \"kubernetes.io/projected/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-kube-api-access-9c4lj\") pod \"ceilometer-0\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " pod="openstack/ceilometer-0" Dec 06 08:31:06 crc kubenswrapper[4945]: I1206 08:31:06.572600 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 08:31:07 crc kubenswrapper[4945]: I1206 08:31:07.013072 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbe0274d-42a7-4730-ae93-3fbb07e6e8c8" path="/var/lib/kubelet/pods/bbe0274d-42a7-4730-ae93-3fbb07e6e8c8/volumes" Dec 06 08:31:07 crc kubenswrapper[4945]: I1206 08:31:07.493073 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:07 crc kubenswrapper[4945]: I1206 08:31:07.860076 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d38018-5cbb-4a61-bc47-1b773bbeb0c8","Type":"ContainerStarted","Data":"f71f13eb5ca45c48ebe7ac28684e92f97bb19be2b8e14cf369308b6204c7e92a"} Dec 06 08:31:08 crc kubenswrapper[4945]: I1206 08:31:08.870697 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d38018-5cbb-4a61-bc47-1b773bbeb0c8","Type":"ContainerStarted","Data":"4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42"} Dec 06 08:31:08 crc kubenswrapper[4945]: I1206 08:31:08.870994 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d38018-5cbb-4a61-bc47-1b773bbeb0c8","Type":"ContainerStarted","Data":"f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728"} Dec 06 08:31:09 crc kubenswrapper[4945]: I1206 08:31:09.887912 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d38018-5cbb-4a61-bc47-1b773bbeb0c8","Type":"ContainerStarted","Data":"e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460"} Dec 06 08:31:10 crc kubenswrapper[4945]: I1206 08:31:10.899689 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d38018-5cbb-4a61-bc47-1b773bbeb0c8","Type":"ContainerStarted","Data":"cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1"} Dec 06 08:31:10 crc kubenswrapper[4945]: I1206 08:31:10.901473 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 08:31:10 crc kubenswrapper[4945]: I1206 08:31:10.929630 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.188480695 podStartE2EDuration="4.929610623s" podCreationTimestamp="2025-12-06 08:31:06 +0000 UTC" firstStartedPulling="2025-12-06 08:31:07.495116544 +0000 UTC m=+5880.949977588" lastFinishedPulling="2025-12-06 08:31:10.236246462 +0000 UTC m=+5883.691107516" observedRunningTime="2025-12-06 08:31:10.92801161 +0000 UTC m=+5884.382872654" watchObservedRunningTime="2025-12-06 08:31:10.929610623 +0000 UTC m=+5884.384471657" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.194636 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-t4zbd"] Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.196258 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t4zbd" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.226555 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-t4zbd"] Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.296155 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-73a0-account-create-update-4fjf2"] Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.297897 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-73a0-account-create-update-4fjf2" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.300547 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.309751 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-73a0-account-create-update-4fjf2"] Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.317774 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2170110e-abfb-4338-bc65-9b2e244f6a90-operator-scripts\") pod \"manila-db-create-t4zbd\" (UID: \"2170110e-abfb-4338-bc65-9b2e244f6a90\") " pod="openstack/manila-db-create-t4zbd" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.317993 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtzxj\" (UniqueName: \"kubernetes.io/projected/2170110e-abfb-4338-bc65-9b2e244f6a90-kube-api-access-dtzxj\") pod \"manila-db-create-t4zbd\" (UID: \"2170110e-abfb-4338-bc65-9b2e244f6a90\") " pod="openstack/manila-db-create-t4zbd" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.420416 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2170110e-abfb-4338-bc65-9b2e244f6a90-operator-scripts\") pod \"manila-db-create-t4zbd\" (UID: \"2170110e-abfb-4338-bc65-9b2e244f6a90\") " pod="openstack/manila-db-create-t4zbd" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.420544 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njzzv\" (UniqueName: \"kubernetes.io/projected/defcea3b-acfe-4ea2-9080-0b70d3c567f4-kube-api-access-njzzv\") pod \"manila-73a0-account-create-update-4fjf2\" (UID: \"defcea3b-acfe-4ea2-9080-0b70d3c567f4\") " pod="openstack/manila-73a0-account-create-update-4fjf2" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.420621 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtzxj\" (UniqueName: \"kubernetes.io/projected/2170110e-abfb-4338-bc65-9b2e244f6a90-kube-api-access-dtzxj\") pod \"manila-db-create-t4zbd\" (UID: \"2170110e-abfb-4338-bc65-9b2e244f6a90\") " pod="openstack/manila-db-create-t4zbd" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.420688 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/defcea3b-acfe-4ea2-9080-0b70d3c567f4-operator-scripts\") pod \"manila-73a0-account-create-update-4fjf2\" (UID: \"defcea3b-acfe-4ea2-9080-0b70d3c567f4\") " pod="openstack/manila-73a0-account-create-update-4fjf2" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.421958 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2170110e-abfb-4338-bc65-9b2e244f6a90-operator-scripts\") pod \"manila-db-create-t4zbd\" (UID: \"2170110e-abfb-4338-bc65-9b2e244f6a90\") " pod="openstack/manila-db-create-t4zbd" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.467986 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtzxj\" (UniqueName: \"kubernetes.io/projected/2170110e-abfb-4338-bc65-9b2e244f6a90-kube-api-access-dtzxj\") pod \"manila-db-create-t4zbd\" (UID: \"2170110e-abfb-4338-bc65-9b2e244f6a90\") " pod="openstack/manila-db-create-t4zbd" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.519935 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t4zbd" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.521863 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njzzv\" (UniqueName: \"kubernetes.io/projected/defcea3b-acfe-4ea2-9080-0b70d3c567f4-kube-api-access-njzzv\") pod \"manila-73a0-account-create-update-4fjf2\" (UID: \"defcea3b-acfe-4ea2-9080-0b70d3c567f4\") " pod="openstack/manila-73a0-account-create-update-4fjf2" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.522027 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/defcea3b-acfe-4ea2-9080-0b70d3c567f4-operator-scripts\") pod \"manila-73a0-account-create-update-4fjf2\" (UID: \"defcea3b-acfe-4ea2-9080-0b70d3c567f4\") " pod="openstack/manila-73a0-account-create-update-4fjf2" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.522951 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/defcea3b-acfe-4ea2-9080-0b70d3c567f4-operator-scripts\") pod \"manila-73a0-account-create-update-4fjf2\" (UID: \"defcea3b-acfe-4ea2-9080-0b70d3c567f4\") " pod="openstack/manila-73a0-account-create-update-4fjf2" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.571877 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njzzv\" (UniqueName: \"kubernetes.io/projected/defcea3b-acfe-4ea2-9080-0b70d3c567f4-kube-api-access-njzzv\") pod \"manila-73a0-account-create-update-4fjf2\" (UID: \"defcea3b-acfe-4ea2-9080-0b70d3c567f4\") " pod="openstack/manila-73a0-account-create-update-4fjf2" Dec 06 08:31:11 crc kubenswrapper[4945]: I1206 08:31:11.617473 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-73a0-account-create-update-4fjf2" Dec 06 08:31:12 crc kubenswrapper[4945]: I1206 08:31:12.163915 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-t4zbd"] Dec 06 08:31:12 crc kubenswrapper[4945]: I1206 08:31:12.268879 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-73a0-account-create-update-4fjf2"] Dec 06 08:31:12 crc kubenswrapper[4945]: W1206 08:31:12.275168 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddefcea3b_acfe_4ea2_9080_0b70d3c567f4.slice/crio-b897dba723770cd77ca3e44e42992e43831082921a1e5aef13c1800763c9b8fb WatchSource:0}: Error finding container b897dba723770cd77ca3e44e42992e43831082921a1e5aef13c1800763c9b8fb: Status 404 returned error can't find the container with id b897dba723770cd77ca3e44e42992e43831082921a1e5aef13c1800763c9b8fb Dec 06 08:31:12 crc kubenswrapper[4945]: I1206 08:31:12.917919 4945 generic.go:334] "Generic (PLEG): container finished" podID="2170110e-abfb-4338-bc65-9b2e244f6a90" containerID="a193e974e0a1534ae2f0e810dd25b74b7902c0abf069fe050356abdd47732aa8" exitCode=0 Dec 06 08:31:12 crc kubenswrapper[4945]: I1206 08:31:12.918056 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-t4zbd" event={"ID":"2170110e-abfb-4338-bc65-9b2e244f6a90","Type":"ContainerDied","Data":"a193e974e0a1534ae2f0e810dd25b74b7902c0abf069fe050356abdd47732aa8"} Dec 06 08:31:12 crc kubenswrapper[4945]: I1206 08:31:12.918244 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-t4zbd" event={"ID":"2170110e-abfb-4338-bc65-9b2e244f6a90","Type":"ContainerStarted","Data":"76b6496f8b807509e23ed1c7c776688540d2437779ef30742f930b24d86820c9"} Dec 06 08:31:12 crc kubenswrapper[4945]: I1206 08:31:12.941331 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-73a0-account-create-update-4fjf2" event={"ID":"defcea3b-acfe-4ea2-9080-0b70d3c567f4","Type":"ContainerStarted","Data":"28825cdd92c87d215b303d6224f8cb359bd70f1263383b50e4ab7003291b9313"} Dec 06 08:31:12 crc kubenswrapper[4945]: I1206 08:31:12.941499 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-73a0-account-create-update-4fjf2" event={"ID":"defcea3b-acfe-4ea2-9080-0b70d3c567f4","Type":"ContainerStarted","Data":"b897dba723770cd77ca3e44e42992e43831082921a1e5aef13c1800763c9b8fb"} Dec 06 08:31:12 crc kubenswrapper[4945]: I1206 08:31:12.979061 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-73a0-account-create-update-4fjf2" podStartSLOduration=1.9790411940000001 podStartE2EDuration="1.979041194s" podCreationTimestamp="2025-12-06 08:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:31:12.962733488 +0000 UTC m=+5886.417594542" watchObservedRunningTime="2025-12-06 08:31:12.979041194 +0000 UTC m=+5886.433902228" Dec 06 08:31:13 crc kubenswrapper[4945]: I1206 08:31:13.041309 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-5tkvt"] Dec 06 08:31:13 crc kubenswrapper[4945]: I1206 08:31:13.061014 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-5tkvt"] Dec 06 08:31:13 crc kubenswrapper[4945]: I1206 08:31:13.949205 4945 generic.go:334] "Generic (PLEG): container finished" podID="defcea3b-acfe-4ea2-9080-0b70d3c567f4" containerID="28825cdd92c87d215b303d6224f8cb359bd70f1263383b50e4ab7003291b9313" exitCode=0 Dec 06 08:31:13 crc kubenswrapper[4945]: I1206 08:31:13.949310 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-73a0-account-create-update-4fjf2" event={"ID":"defcea3b-acfe-4ea2-9080-0b70d3c567f4","Type":"ContainerDied","Data":"28825cdd92c87d215b303d6224f8cb359bd70f1263383b50e4ab7003291b9313"} Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.347493 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t4zbd" Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.494912 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2170110e-abfb-4338-bc65-9b2e244f6a90-operator-scripts\") pod \"2170110e-abfb-4338-bc65-9b2e244f6a90\" (UID: \"2170110e-abfb-4338-bc65-9b2e244f6a90\") " Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.495191 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtzxj\" (UniqueName: \"kubernetes.io/projected/2170110e-abfb-4338-bc65-9b2e244f6a90-kube-api-access-dtzxj\") pod \"2170110e-abfb-4338-bc65-9b2e244f6a90\" (UID: \"2170110e-abfb-4338-bc65-9b2e244f6a90\") " Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.495480 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2170110e-abfb-4338-bc65-9b2e244f6a90-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2170110e-abfb-4338-bc65-9b2e244f6a90" (UID: "2170110e-abfb-4338-bc65-9b2e244f6a90"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.496124 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2170110e-abfb-4338-bc65-9b2e244f6a90-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.499964 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2170110e-abfb-4338-bc65-9b2e244f6a90-kube-api-access-dtzxj" (OuterVolumeSpecName: "kube-api-access-dtzxj") pod "2170110e-abfb-4338-bc65-9b2e244f6a90" (UID: "2170110e-abfb-4338-bc65-9b2e244f6a90"). InnerVolumeSpecName "kube-api-access-dtzxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.598065 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtzxj\" (UniqueName: \"kubernetes.io/projected/2170110e-abfb-4338-bc65-9b2e244f6a90-kube-api-access-dtzxj\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.965873 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71bda92d-e6dd-4061-874f-4020f86a223e" path="/var/lib/kubelet/pods/71bda92d-e6dd-4061-874f-4020f86a223e/volumes" Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.968596 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-t4zbd" event={"ID":"2170110e-abfb-4338-bc65-9b2e244f6a90","Type":"ContainerDied","Data":"76b6496f8b807509e23ed1c7c776688540d2437779ef30742f930b24d86820c9"} Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.968632 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t4zbd" Dec 06 08:31:14 crc kubenswrapper[4945]: I1206 08:31:14.968638 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76b6496f8b807509e23ed1c7c776688540d2437779ef30742f930b24d86820c9" Dec 06 08:31:15 crc kubenswrapper[4945]: I1206 08:31:15.487008 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-73a0-account-create-update-4fjf2" Dec 06 08:31:15 crc kubenswrapper[4945]: I1206 08:31:15.644573 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/defcea3b-acfe-4ea2-9080-0b70d3c567f4-operator-scripts\") pod \"defcea3b-acfe-4ea2-9080-0b70d3c567f4\" (UID: \"defcea3b-acfe-4ea2-9080-0b70d3c567f4\") " Dec 06 08:31:15 crc kubenswrapper[4945]: I1206 08:31:15.644777 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njzzv\" (UniqueName: \"kubernetes.io/projected/defcea3b-acfe-4ea2-9080-0b70d3c567f4-kube-api-access-njzzv\") pod \"defcea3b-acfe-4ea2-9080-0b70d3c567f4\" (UID: \"defcea3b-acfe-4ea2-9080-0b70d3c567f4\") " Dec 06 08:31:15 crc kubenswrapper[4945]: I1206 08:31:15.645140 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/defcea3b-acfe-4ea2-9080-0b70d3c567f4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "defcea3b-acfe-4ea2-9080-0b70d3c567f4" (UID: "defcea3b-acfe-4ea2-9080-0b70d3c567f4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:31:15 crc kubenswrapper[4945]: I1206 08:31:15.650615 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/defcea3b-acfe-4ea2-9080-0b70d3c567f4-kube-api-access-njzzv" (OuterVolumeSpecName: "kube-api-access-njzzv") pod "defcea3b-acfe-4ea2-9080-0b70d3c567f4" (UID: "defcea3b-acfe-4ea2-9080-0b70d3c567f4"). InnerVolumeSpecName "kube-api-access-njzzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:31:15 crc kubenswrapper[4945]: I1206 08:31:15.746903 4945 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/defcea3b-acfe-4ea2-9080-0b70d3c567f4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:15 crc kubenswrapper[4945]: I1206 08:31:15.746936 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njzzv\" (UniqueName: \"kubernetes.io/projected/defcea3b-acfe-4ea2-9080-0b70d3c567f4-kube-api-access-njzzv\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:15 crc kubenswrapper[4945]: I1206 08:31:15.978460 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-73a0-account-create-update-4fjf2" event={"ID":"defcea3b-acfe-4ea2-9080-0b70d3c567f4","Type":"ContainerDied","Data":"b897dba723770cd77ca3e44e42992e43831082921a1e5aef13c1800763c9b8fb"} Dec 06 08:31:15 crc kubenswrapper[4945]: I1206 08:31:15.978709 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b897dba723770cd77ca3e44e42992e43831082921a1e5aef13c1800763c9b8fb" Dec 06 08:31:15 crc kubenswrapper[4945]: I1206 08:31:15.978512 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-73a0-account-create-update-4fjf2" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.536446 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-gq7rn"] Dec 06 08:31:16 crc kubenswrapper[4945]: E1206 08:31:16.536861 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2170110e-abfb-4338-bc65-9b2e244f6a90" containerName="mariadb-database-create" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.536880 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2170110e-abfb-4338-bc65-9b2e244f6a90" containerName="mariadb-database-create" Dec 06 08:31:16 crc kubenswrapper[4945]: E1206 08:31:16.536897 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="defcea3b-acfe-4ea2-9080-0b70d3c567f4" containerName="mariadb-account-create-update" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.536903 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="defcea3b-acfe-4ea2-9080-0b70d3c567f4" containerName="mariadb-account-create-update" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.537131 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="defcea3b-acfe-4ea2-9080-0b70d3c567f4" containerName="mariadb-account-create-update" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.537150 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2170110e-abfb-4338-bc65-9b2e244f6a90" containerName="mariadb-database-create" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.537830 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.540624 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-jdp74" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.544564 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.564459 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-gq7rn"] Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.669413 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-config-data\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.669573 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-combined-ca-bundle\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.669609 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-job-config-data\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.669630 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltm6t\" (UniqueName: \"kubernetes.io/projected/81111732-b6e5-4c98-8372-58f18db7ed17-kube-api-access-ltm6t\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.772086 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-combined-ca-bundle\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.772150 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-job-config-data\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.772178 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltm6t\" (UniqueName: \"kubernetes.io/projected/81111732-b6e5-4c98-8372-58f18db7ed17-kube-api-access-ltm6t\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.772309 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-config-data\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.777426 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-combined-ca-bundle\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.777826 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-job-config-data\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.785038 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-config-data\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.792720 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltm6t\" (UniqueName: \"kubernetes.io/projected/81111732-b6e5-4c98-8372-58f18db7ed17-kube-api-access-ltm6t\") pod \"manila-db-sync-gq7rn\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:16 crc kubenswrapper[4945]: I1206 08:31:16.857016 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:17 crc kubenswrapper[4945]: I1206 08:31:17.597721 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-gq7rn"] Dec 06 08:31:18 crc kubenswrapper[4945]: I1206 08:31:18.004890 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-gq7rn" event={"ID":"81111732-b6e5-4c98-8372-58f18db7ed17","Type":"ContainerStarted","Data":"9fdf06bd3e19a6e0081f6b35cad9b864a368b7b90884177d54379e47bb51bda1"} Dec 06 08:31:24 crc kubenswrapper[4945]: I1206 08:31:24.061707 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-gq7rn" event={"ID":"81111732-b6e5-4c98-8372-58f18db7ed17","Type":"ContainerStarted","Data":"360b3ab8f0b3479da2025dd9eb9ac1d9ad5e162b412d76dfb351e5c4418e31a9"} Dec 06 08:31:24 crc kubenswrapper[4945]: I1206 08:31:24.085665 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-gq7rn" podStartSLOduration=2.530917134 podStartE2EDuration="8.085645839s" podCreationTimestamp="2025-12-06 08:31:16 +0000 UTC" firstStartedPulling="2025-12-06 08:31:17.606135111 +0000 UTC m=+5891.060996155" lastFinishedPulling="2025-12-06 08:31:23.160863816 +0000 UTC m=+5896.615724860" observedRunningTime="2025-12-06 08:31:24.073398541 +0000 UTC m=+5897.528259585" watchObservedRunningTime="2025-12-06 08:31:24.085645839 +0000 UTC m=+5897.540506883" Dec 06 08:31:26 crc kubenswrapper[4945]: I1206 08:31:26.081420 4945 generic.go:334] "Generic (PLEG): container finished" podID="81111732-b6e5-4c98-8372-58f18db7ed17" containerID="360b3ab8f0b3479da2025dd9eb9ac1d9ad5e162b412d76dfb351e5c4418e31a9" exitCode=0 Dec 06 08:31:26 crc kubenswrapper[4945]: I1206 08:31:26.081541 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-gq7rn" event={"ID":"81111732-b6e5-4c98-8372-58f18db7ed17","Type":"ContainerDied","Data":"360b3ab8f0b3479da2025dd9eb9ac1d9ad5e162b412d76dfb351e5c4418e31a9"} Dec 06 08:31:26 crc kubenswrapper[4945]: I1206 08:31:26.587684 4945 scope.go:117] "RemoveContainer" containerID="6c6a9a23d4d92392bb44d53ab56d1faff17e9683c0970184f7b530749924d9cf" Dec 06 08:31:26 crc kubenswrapper[4945]: I1206 08:31:26.628690 4945 scope.go:117] "RemoveContainer" containerID="3eb2f8ee314c0830aadbbffe9617798c35e65645322e58766718d24477a80b16" Dec 06 08:31:26 crc kubenswrapper[4945]: I1206 08:31:26.666990 4945 scope.go:117] "RemoveContainer" containerID="d8e46074d4c0283a0851990cddde3645febfde2fd6047ae34a343c449cd2ef1f" Dec 06 08:31:26 crc kubenswrapper[4945]: I1206 08:31:26.723232 4945 scope.go:117] "RemoveContainer" containerID="943434f63afced1af6932d562efb514a4cadbba4a237a0e7424949164d1e9fb9" Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.493970 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.580176 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-job-config-data\") pod \"81111732-b6e5-4c98-8372-58f18db7ed17\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.580736 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-config-data\") pod \"81111732-b6e5-4c98-8372-58f18db7ed17\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.580775 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-combined-ca-bundle\") pod \"81111732-b6e5-4c98-8372-58f18db7ed17\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.580950 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltm6t\" (UniqueName: \"kubernetes.io/projected/81111732-b6e5-4c98-8372-58f18db7ed17-kube-api-access-ltm6t\") pod \"81111732-b6e5-4c98-8372-58f18db7ed17\" (UID: \"81111732-b6e5-4c98-8372-58f18db7ed17\") " Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.586999 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "81111732-b6e5-4c98-8372-58f18db7ed17" (UID: "81111732-b6e5-4c98-8372-58f18db7ed17"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.587105 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81111732-b6e5-4c98-8372-58f18db7ed17-kube-api-access-ltm6t" (OuterVolumeSpecName: "kube-api-access-ltm6t") pod "81111732-b6e5-4c98-8372-58f18db7ed17" (UID: "81111732-b6e5-4c98-8372-58f18db7ed17"). InnerVolumeSpecName "kube-api-access-ltm6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.590493 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-config-data" (OuterVolumeSpecName: "config-data") pod "81111732-b6e5-4c98-8372-58f18db7ed17" (UID: "81111732-b6e5-4c98-8372-58f18db7ed17"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.613524 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81111732-b6e5-4c98-8372-58f18db7ed17" (UID: "81111732-b6e5-4c98-8372-58f18db7ed17"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.687097 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.687142 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.687158 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltm6t\" (UniqueName: \"kubernetes.io/projected/81111732-b6e5-4c98-8372-58f18db7ed17-kube-api-access-ltm6t\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:27 crc kubenswrapper[4945]: I1206 08:31:27.687170 4945 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/81111732-b6e5-4c98-8372-58f18db7ed17-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.102036 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-gq7rn" event={"ID":"81111732-b6e5-4c98-8372-58f18db7ed17","Type":"ContainerDied","Data":"9fdf06bd3e19a6e0081f6b35cad9b864a368b7b90884177d54379e47bb51bda1"} Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.102069 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fdf06bd3e19a6e0081f6b35cad9b864a368b7b90884177d54379e47bb51bda1" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.102087 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-gq7rn" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.355939 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 06 08:31:28 crc kubenswrapper[4945]: E1206 08:31:28.356459 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81111732-b6e5-4c98-8372-58f18db7ed17" containerName="manila-db-sync" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.356485 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="81111732-b6e5-4c98-8372-58f18db7ed17" containerName="manila-db-sync" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.356728 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="81111732-b6e5-4c98-8372-58f18db7ed17" containerName="manila-db-sync" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.358162 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.376072 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.377973 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.378126 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-jdp74" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.378255 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.407487 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.502987 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.503055 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-scripts\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.503098 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c72bs\" (UniqueName: \"kubernetes.io/projected/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-kube-api-access-c72bs\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.503134 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.503192 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-config-data\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.503322 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.529261 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.532071 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.537089 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.558533 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.646747 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647067 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647127 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-config-data\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647155 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647189 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stk7h\" (UniqueName: \"kubernetes.io/projected/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-kube-api-access-stk7h\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647214 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-ceph\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647268 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647309 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-scripts\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647363 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c72bs\" (UniqueName: \"kubernetes.io/projected/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-kube-api-access-c72bs\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647389 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647469 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647506 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-config-data\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647565 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.647604 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-scripts\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.648841 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.671270 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-config-data\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.682812 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c72bs\" (UniqueName: \"kubernetes.io/projected/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-kube-api-access-c72bs\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.685345 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.687588 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.698961 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f-scripts\") pod \"manila-scheduler-0\" (UID: \"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f\") " pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.706734 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67849f8f7f-fbcsq"] Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.710268 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.759880 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67849f8f7f-fbcsq"] Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.761228 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.761918 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-config-data\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.762034 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stk7h\" (UniqueName: \"kubernetes.io/projected/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-kube-api-access-stk7h\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.762102 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-ceph\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.762379 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.763476 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.763592 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.763746 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-scripts\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.769809 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.769987 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.770380 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-scripts\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.788850 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.800867 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.801815 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-ceph\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.809261 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stk7h\" (UniqueName: \"kubernetes.io/projected/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-kube-api-access-stk7h\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.812065 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b19fcd2e-e77a-4543-ac62-5ecef9e48b7c-config-data\") pod \"manila-share-share1-0\" (UID: \"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c\") " pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.866577 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-config\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.866624 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxxlf\" (UniqueName: \"kubernetes.io/projected/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-kube-api-access-rxxlf\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.866695 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-sb\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.866750 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-nb\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.866771 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-dns-svc\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.874635 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.927775 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.929582 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.944225 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.975147 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.987869 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.989321 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-nb\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.989401 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-dns-svc\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.989535 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-config\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.989569 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxxlf\" (UniqueName: \"kubernetes.io/projected/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-kube-api-access-rxxlf\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.989673 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-sb\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.990248 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-nb\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.990616 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-sb\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.990881 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-config\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:28 crc kubenswrapper[4945]: I1206 08:31:28.991272 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-dns-svc\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.026314 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxxlf\" (UniqueName: \"kubernetes.io/projected/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-kube-api-access-rxxlf\") pod \"dnsmasq-dns-67849f8f7f-fbcsq\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.096491 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c07d97d9-2707-4564-86b5-f699f80d5831-etc-machine-id\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.096725 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c07d97d9-2707-4564-86b5-f699f80d5831-logs\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.096762 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-scripts\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.096781 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58bc6\" (UniqueName: \"kubernetes.io/projected/c07d97d9-2707-4564-86b5-f699f80d5831-kube-api-access-58bc6\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.096795 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.096836 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-config-data\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.096952 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-config-data-custom\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.199947 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.200996 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-config-data-custom\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.201108 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c07d97d9-2707-4564-86b5-f699f80d5831-etc-machine-id\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.201144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c07d97d9-2707-4564-86b5-f699f80d5831-logs\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.201193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58bc6\" (UniqueName: \"kubernetes.io/projected/c07d97d9-2707-4564-86b5-f699f80d5831-kube-api-access-58bc6\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.201215 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-scripts\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.201241 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.201321 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-config-data\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.201977 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c07d97d9-2707-4564-86b5-f699f80d5831-logs\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.206064 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c07d97d9-2707-4564-86b5-f699f80d5831-etc-machine-id\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.207757 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-config-data-custom\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.212129 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.216129 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-scripts\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.224935 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c07d97d9-2707-4564-86b5-f699f80d5831-config-data\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.236415 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58bc6\" (UniqueName: \"kubernetes.io/projected/c07d97d9-2707-4564-86b5-f699f80d5831-kube-api-access-58bc6\") pod \"manila-api-0\" (UID: \"c07d97d9-2707-4564-86b5-f699f80d5831\") " pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.314334 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.660246 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.680953 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 06 08:31:29 crc kubenswrapper[4945]: I1206 08:31:29.798557 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67849f8f7f-fbcsq"] Dec 06 08:31:30 crc kubenswrapper[4945]: I1206 08:31:30.118329 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 06 08:31:30 crc kubenswrapper[4945]: W1206 08:31:30.145225 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc07d97d9_2707_4564_86b5_f699f80d5831.slice/crio-a989923b2c8f54138c0d5cc3ff2cd4104025c8fcf9671a2a8518072f489a08ec WatchSource:0}: Error finding container a989923b2c8f54138c0d5cc3ff2cd4104025c8fcf9671a2a8518072f489a08ec: Status 404 returned error can't find the container with id a989923b2c8f54138c0d5cc3ff2cd4104025c8fcf9671a2a8518072f489a08ec Dec 06 08:31:30 crc kubenswrapper[4945]: I1206 08:31:30.165732 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" event={"ID":"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5","Type":"ContainerStarted","Data":"39036e42258c5cbffa6b6954ae05fc139603ed07839d1d6acc3734d7466bf1f4"} Dec 06 08:31:30 crc kubenswrapper[4945]: I1206 08:31:30.168561 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c","Type":"ContainerStarted","Data":"a09573b1f71581048f8eaaf7a84795126e4f0548b9212ec3bc0bbbd5fbe0e0e4"} Dec 06 08:31:30 crc kubenswrapper[4945]: I1206 08:31:30.181144 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f","Type":"ContainerStarted","Data":"495a1702a71bb0b7fba305cd043170a0e91cb1d2d9f923f7161ad56a12a4eedd"} Dec 06 08:31:31 crc kubenswrapper[4945]: I1206 08:31:31.207136 4945 generic.go:334] "Generic (PLEG): container finished" podID="91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" containerID="a6db0fbef174650b10873f487029094f8b7e172c43be62c0133bea48ffacc4c6" exitCode=0 Dec 06 08:31:31 crc kubenswrapper[4945]: I1206 08:31:31.208153 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" event={"ID":"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5","Type":"ContainerDied","Data":"a6db0fbef174650b10873f487029094f8b7e172c43be62c0133bea48ffacc4c6"} Dec 06 08:31:31 crc kubenswrapper[4945]: I1206 08:31:31.222676 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c07d97d9-2707-4564-86b5-f699f80d5831","Type":"ContainerStarted","Data":"7c12407a7c703bfe38ead7d5c29cf6348ce2936b495599d842f9d808c5f7d7b9"} Dec 06 08:31:31 crc kubenswrapper[4945]: I1206 08:31:31.222904 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c07d97d9-2707-4564-86b5-f699f80d5831","Type":"ContainerStarted","Data":"a989923b2c8f54138c0d5cc3ff2cd4104025c8fcf9671a2a8518072f489a08ec"} Dec 06 08:31:31 crc kubenswrapper[4945]: I1206 08:31:31.243408 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f","Type":"ContainerStarted","Data":"19560ad51692d693d1401e07e004196524679fa486107813defd3b7c381a4432"} Dec 06 08:31:32 crc kubenswrapper[4945]: I1206 08:31:32.266662 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c07d97d9-2707-4564-86b5-f699f80d5831","Type":"ContainerStarted","Data":"e014bd970f55072e13f6412a2ad327208919e617073dbaba23f9fb6c3478f73a"} Dec 06 08:31:32 crc kubenswrapper[4945]: I1206 08:31:32.268079 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 06 08:31:32 crc kubenswrapper[4945]: I1206 08:31:32.269902 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f","Type":"ContainerStarted","Data":"113a3880b556776708b86ddef2bcac93feb74c72b004e655123c732cd3d01e08"} Dec 06 08:31:32 crc kubenswrapper[4945]: I1206 08:31:32.272793 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" event={"ID":"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5","Type":"ContainerStarted","Data":"66b1e89d0ae3f4fe048dc9563f39ffe0c4521bb95900bcf8d878792652e5fbbc"} Dec 06 08:31:32 crc kubenswrapper[4945]: I1206 08:31:32.272963 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:32 crc kubenswrapper[4945]: I1206 08:31:32.308103 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.308081789 podStartE2EDuration="4.308081789s" podCreationTimestamp="2025-12-06 08:31:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:31:32.292873912 +0000 UTC m=+5905.747734956" watchObservedRunningTime="2025-12-06 08:31:32.308081789 +0000 UTC m=+5905.762942833" Dec 06 08:31:32 crc kubenswrapper[4945]: I1206 08:31:32.327787 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" podStartSLOduration=4.327764965 podStartE2EDuration="4.327764965s" podCreationTimestamp="2025-12-06 08:31:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:31:32.318242 +0000 UTC m=+5905.773103044" watchObservedRunningTime="2025-12-06 08:31:32.327764965 +0000 UTC m=+5905.782626009" Dec 06 08:31:32 crc kubenswrapper[4945]: I1206 08:31:32.348161 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.949583897 podStartE2EDuration="4.34814225s" podCreationTimestamp="2025-12-06 08:31:28 +0000 UTC" firstStartedPulling="2025-12-06 08:31:29.682653575 +0000 UTC m=+5903.137514619" lastFinishedPulling="2025-12-06 08:31:30.081211928 +0000 UTC m=+5903.536072972" observedRunningTime="2025-12-06 08:31:32.342416777 +0000 UTC m=+5905.797277821" watchObservedRunningTime="2025-12-06 08:31:32.34814225 +0000 UTC m=+5905.803003294" Dec 06 08:31:36 crc kubenswrapper[4945]: I1206 08:31:36.582847 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 08:31:38 crc kubenswrapper[4945]: I1206 08:31:38.989250 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 06 08:31:39 crc kubenswrapper[4945]: I1206 08:31:39.211876 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:31:39 crc kubenswrapper[4945]: I1206 08:31:39.279504 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64d4b844c-l5f9k"] Dec 06 08:31:39 crc kubenswrapper[4945]: I1206 08:31:39.279740 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" podUID="b1e39b0c-dfae-455c-9947-7af9b39a45f9" containerName="dnsmasq-dns" containerID="cri-o://f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c" gracePeriod=10 Dec 06 08:31:39 crc kubenswrapper[4945]: I1206 08:31:39.399831 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c","Type":"ContainerStarted","Data":"a8e0e64e13d8a1ca6550bc3d15ec89139cb14f71b45be69365808bb06f12efc1"} Dec 06 08:31:39 crc kubenswrapper[4945]: I1206 08:31:39.408978 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" podUID="b1e39b0c-dfae-455c-9947-7af9b39a45f9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.76:5353: connect: connection refused" Dec 06 08:31:39 crc kubenswrapper[4945]: I1206 08:31:39.912442 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.091071 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-dns-svc\") pod \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.091238 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-config\") pod \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.091302 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-nb\") pod \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.091439 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-sb\") pod \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.091469 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlw87\" (UniqueName: \"kubernetes.io/projected/b1e39b0c-dfae-455c-9947-7af9b39a45f9-kube-api-access-jlw87\") pod \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\" (UID: \"b1e39b0c-dfae-455c-9947-7af9b39a45f9\") " Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.112427 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1e39b0c-dfae-455c-9947-7af9b39a45f9-kube-api-access-jlw87" (OuterVolumeSpecName: "kube-api-access-jlw87") pod "b1e39b0c-dfae-455c-9947-7af9b39a45f9" (UID: "b1e39b0c-dfae-455c-9947-7af9b39a45f9"). InnerVolumeSpecName "kube-api-access-jlw87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.146847 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b1e39b0c-dfae-455c-9947-7af9b39a45f9" (UID: "b1e39b0c-dfae-455c-9947-7af9b39a45f9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.153704 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b1e39b0c-dfae-455c-9947-7af9b39a45f9" (UID: "b1e39b0c-dfae-455c-9947-7af9b39a45f9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.165729 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-config" (OuterVolumeSpecName: "config") pod "b1e39b0c-dfae-455c-9947-7af9b39a45f9" (UID: "b1e39b0c-dfae-455c-9947-7af9b39a45f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.173031 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b1e39b0c-dfae-455c-9947-7af9b39a45f9" (UID: "b1e39b0c-dfae-455c-9947-7af9b39a45f9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.194353 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.194384 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.194400 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.194412 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlw87\" (UniqueName: \"kubernetes.io/projected/b1e39b0c-dfae-455c-9947-7af9b39a45f9-kube-api-access-jlw87\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.194423 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1e39b0c-dfae-455c-9947-7af9b39a45f9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.411377 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b19fcd2e-e77a-4543-ac62-5ecef9e48b7c","Type":"ContainerStarted","Data":"2b8df33d7cc40271c7f35e54aa91d6c38a49b21b92ab2bbfabfad7a0e057ae1d"} Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.414660 4945 generic.go:334] "Generic (PLEG): container finished" podID="b1e39b0c-dfae-455c-9947-7af9b39a45f9" containerID="f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c" exitCode=0 Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.414699 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" event={"ID":"b1e39b0c-dfae-455c-9947-7af9b39a45f9","Type":"ContainerDied","Data":"f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c"} Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.414723 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" event={"ID":"b1e39b0c-dfae-455c-9947-7af9b39a45f9","Type":"ContainerDied","Data":"ebad84242f34f71d444cacd337c05c8be6560872a9b518550cbdc7d604438bd2"} Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.414740 4945 scope.go:117] "RemoveContainer" containerID="f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.414803 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d4b844c-l5f9k" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.434883 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.723779114 podStartE2EDuration="12.434867008s" podCreationTimestamp="2025-12-06 08:31:28 +0000 UTC" firstStartedPulling="2025-12-06 08:31:29.666475512 +0000 UTC m=+5903.121336556" lastFinishedPulling="2025-12-06 08:31:38.377563396 +0000 UTC m=+5911.832424450" observedRunningTime="2025-12-06 08:31:40.431557159 +0000 UTC m=+5913.886418213" watchObservedRunningTime="2025-12-06 08:31:40.434867008 +0000 UTC m=+5913.889728042" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.455564 4945 scope.go:117] "RemoveContainer" containerID="8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.488436 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64d4b844c-l5f9k"] Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.491537 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64d4b844c-l5f9k"] Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.498711 4945 scope.go:117] "RemoveContainer" containerID="f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c" Dec 06 08:31:40 crc kubenswrapper[4945]: E1206 08:31:40.499218 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c\": container with ID starting with f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c not found: ID does not exist" containerID="f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.499261 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c"} err="failed to get container status \"f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c\": rpc error: code = NotFound desc = could not find container \"f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c\": container with ID starting with f92df1f948d5399d94f94cdea38bdb2a5d51f815e503e80d9db3e472429a7f5c not found: ID does not exist" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.499304 4945 scope.go:117] "RemoveContainer" containerID="8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca" Dec 06 08:31:40 crc kubenswrapper[4945]: E1206 08:31:40.499721 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca\": container with ID starting with 8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca not found: ID does not exist" containerID="8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.499776 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca"} err="failed to get container status \"8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca\": rpc error: code = NotFound desc = could not find container \"8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca\": container with ID starting with 8fd7e719a8ef4c41787cc2e6fa8ac27d445bfb8c9f434492a5f06ec3b0e224ca not found: ID does not exist" Dec 06 08:31:40 crc kubenswrapper[4945]: I1206 08:31:40.974613 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1e39b0c-dfae-455c-9947-7af9b39a45f9" path="/var/lib/kubelet/pods/b1e39b0c-dfae-455c-9947-7af9b39a45f9/volumes" Dec 06 08:31:41 crc kubenswrapper[4945]: I1206 08:31:41.967854 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:41 crc kubenswrapper[4945]: I1206 08:31:41.968114 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="ceilometer-central-agent" containerID="cri-o://f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728" gracePeriod=30 Dec 06 08:31:41 crc kubenswrapper[4945]: I1206 08:31:41.968170 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="proxy-httpd" containerID="cri-o://cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1" gracePeriod=30 Dec 06 08:31:41 crc kubenswrapper[4945]: I1206 08:31:41.968208 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="ceilometer-notification-agent" containerID="cri-o://4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42" gracePeriod=30 Dec 06 08:31:41 crc kubenswrapper[4945]: I1206 08:31:41.968183 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="sg-core" containerID="cri-o://e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460" gracePeriod=30 Dec 06 08:31:42 crc kubenswrapper[4945]: I1206 08:31:42.436099 4945 generic.go:334] "Generic (PLEG): container finished" podID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerID="cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1" exitCode=0 Dec 06 08:31:42 crc kubenswrapper[4945]: I1206 08:31:42.436447 4945 generic.go:334] "Generic (PLEG): container finished" podID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerID="e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460" exitCode=2 Dec 06 08:31:42 crc kubenswrapper[4945]: I1206 08:31:42.436149 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d38018-5cbb-4a61-bc47-1b773bbeb0c8","Type":"ContainerDied","Data":"cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1"} Dec 06 08:31:42 crc kubenswrapper[4945]: I1206 08:31:42.436504 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d38018-5cbb-4a61-bc47-1b773bbeb0c8","Type":"ContainerDied","Data":"e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460"} Dec 06 08:31:43 crc kubenswrapper[4945]: I1206 08:31:43.465594 4945 generic.go:334] "Generic (PLEG): container finished" podID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerID="f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728" exitCode=0 Dec 06 08:31:43 crc kubenswrapper[4945]: I1206 08:31:43.465640 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d38018-5cbb-4a61-bc47-1b773bbeb0c8","Type":"ContainerDied","Data":"f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728"} Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.078516 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.212873 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-scripts\") pod \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.213191 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-sg-core-conf-yaml\") pod \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.213338 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-run-httpd\") pod \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.213537 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-combined-ca-bundle\") pod \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.213668 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "76d38018-5cbb-4a61-bc47-1b773bbeb0c8" (UID: "76d38018-5cbb-4a61-bc47-1b773bbeb0c8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.213820 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c4lj\" (UniqueName: \"kubernetes.io/projected/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-kube-api-access-9c4lj\") pod \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.213936 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-config-data\") pod \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.214097 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-log-httpd\") pod \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\" (UID: \"76d38018-5cbb-4a61-bc47-1b773bbeb0c8\") " Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.214445 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "76d38018-5cbb-4a61-bc47-1b773bbeb0c8" (UID: "76d38018-5cbb-4a61-bc47-1b773bbeb0c8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.214869 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.214956 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.218955 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-scripts" (OuterVolumeSpecName: "scripts") pod "76d38018-5cbb-4a61-bc47-1b773bbeb0c8" (UID: "76d38018-5cbb-4a61-bc47-1b773bbeb0c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.219385 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-kube-api-access-9c4lj" (OuterVolumeSpecName: "kube-api-access-9c4lj") pod "76d38018-5cbb-4a61-bc47-1b773bbeb0c8" (UID: "76d38018-5cbb-4a61-bc47-1b773bbeb0c8"). InnerVolumeSpecName "kube-api-access-9c4lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.242308 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "76d38018-5cbb-4a61-bc47-1b773bbeb0c8" (UID: "76d38018-5cbb-4a61-bc47-1b773bbeb0c8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.318577 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.318618 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.318632 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c4lj\" (UniqueName: \"kubernetes.io/projected/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-kube-api-access-9c4lj\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.320802 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76d38018-5cbb-4a61-bc47-1b773bbeb0c8" (UID: "76d38018-5cbb-4a61-bc47-1b773bbeb0c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.333814 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-config-data" (OuterVolumeSpecName: "config-data") pod "76d38018-5cbb-4a61-bc47-1b773bbeb0c8" (UID: "76d38018-5cbb-4a61-bc47-1b773bbeb0c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.420220 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.420269 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d38018-5cbb-4a61-bc47-1b773bbeb0c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.487753 4945 generic.go:334] "Generic (PLEG): container finished" podID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerID="4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42" exitCode=0 Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.487795 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d38018-5cbb-4a61-bc47-1b773bbeb0c8","Type":"ContainerDied","Data":"4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42"} Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.487839 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d38018-5cbb-4a61-bc47-1b773bbeb0c8","Type":"ContainerDied","Data":"f71f13eb5ca45c48ebe7ac28684e92f97bb19be2b8e14cf369308b6204c7e92a"} Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.487857 4945 scope.go:117] "RemoveContainer" containerID="cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.487872 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.511787 4945 scope.go:117] "RemoveContainer" containerID="e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.535591 4945 scope.go:117] "RemoveContainer" containerID="4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.546242 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.561956 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.565491 4945 scope.go:117] "RemoveContainer" containerID="f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.575492 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:45 crc kubenswrapper[4945]: E1206 08:31:45.575978 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="sg-core" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576002 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="sg-core" Dec 06 08:31:45 crc kubenswrapper[4945]: E1206 08:31:45.576039 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="ceilometer-central-agent" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576046 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="ceilometer-central-agent" Dec 06 08:31:45 crc kubenswrapper[4945]: E1206 08:31:45.576069 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e39b0c-dfae-455c-9947-7af9b39a45f9" containerName="init" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576075 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e39b0c-dfae-455c-9947-7af9b39a45f9" containerName="init" Dec 06 08:31:45 crc kubenswrapper[4945]: E1206 08:31:45.576093 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="proxy-httpd" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576098 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="proxy-httpd" Dec 06 08:31:45 crc kubenswrapper[4945]: E1206 08:31:45.576109 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e39b0c-dfae-455c-9947-7af9b39a45f9" containerName="dnsmasq-dns" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576115 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e39b0c-dfae-455c-9947-7af9b39a45f9" containerName="dnsmasq-dns" Dec 06 08:31:45 crc kubenswrapper[4945]: E1206 08:31:45.576130 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="ceilometer-notification-agent" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576137 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="ceilometer-notification-agent" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576331 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="ceilometer-central-agent" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576341 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e39b0c-dfae-455c-9947-7af9b39a45f9" containerName="dnsmasq-dns" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576366 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="sg-core" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576389 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="ceilometer-notification-agent" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.576400 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" containerName="proxy-httpd" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.578575 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.581062 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.581302 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.591226 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.611495 4945 scope.go:117] "RemoveContainer" containerID="cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1" Dec 06 08:31:45 crc kubenswrapper[4945]: E1206 08:31:45.612368 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1\": container with ID starting with cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1 not found: ID does not exist" containerID="cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.612403 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1"} err="failed to get container status \"cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1\": rpc error: code = NotFound desc = could not find container \"cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1\": container with ID starting with cbe7b32c373aa2f72f2d9435a31b4a9684b1076133ab876bec953112598564e1 not found: ID does not exist" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.612424 4945 scope.go:117] "RemoveContainer" containerID="e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460" Dec 06 08:31:45 crc kubenswrapper[4945]: E1206 08:31:45.612780 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460\": container with ID starting with e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460 not found: ID does not exist" containerID="e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.612807 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460"} err="failed to get container status \"e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460\": rpc error: code = NotFound desc = could not find container \"e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460\": container with ID starting with e6ae70e5ab4ec3e6b62fe399bfd3a3731da03cbe976ef073eead27c9d8002460 not found: ID does not exist" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.612823 4945 scope.go:117] "RemoveContainer" containerID="4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42" Dec 06 08:31:45 crc kubenswrapper[4945]: E1206 08:31:45.613041 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42\": container with ID starting with 4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42 not found: ID does not exist" containerID="4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.613060 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42"} err="failed to get container status \"4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42\": rpc error: code = NotFound desc = could not find container \"4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42\": container with ID starting with 4e5228b5fba78f135695c200c76dd78e6dd15eb7d8d4c9a45fb991764467bd42 not found: ID does not exist" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.613074 4945 scope.go:117] "RemoveContainer" containerID="f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728" Dec 06 08:31:45 crc kubenswrapper[4945]: E1206 08:31:45.613247 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728\": container with ID starting with f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728 not found: ID does not exist" containerID="f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.613289 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728"} err="failed to get container status \"f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728\": rpc error: code = NotFound desc = could not find container \"f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728\": container with ID starting with f4ebeef8cb85fb2680aedfa8c60ccc1f14b2cea7a15083510925812c2eea6728 not found: ID does not exist" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.726834 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-config-data\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.726933 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c910732-ba37-4887-9ab3-32c491361007-log-httpd\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.727231 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lksx\" (UniqueName: \"kubernetes.io/projected/4c910732-ba37-4887-9ab3-32c491361007-kube-api-access-2lksx\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.727378 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.727414 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.727450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-scripts\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.727490 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c910732-ba37-4887-9ab3-32c491361007-run-httpd\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.828898 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.829221 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-scripts\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.829259 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c910732-ba37-4887-9ab3-32c491361007-run-httpd\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.829319 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-config-data\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.829372 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c910732-ba37-4887-9ab3-32c491361007-log-httpd\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.829438 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lksx\" (UniqueName: \"kubernetes.io/projected/4c910732-ba37-4887-9ab3-32c491361007-kube-api-access-2lksx\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.829482 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.829929 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c910732-ba37-4887-9ab3-32c491361007-run-httpd\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.830114 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c910732-ba37-4887-9ab3-32c491361007-log-httpd\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.834738 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.835341 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-config-data\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.837779 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-scripts\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.846171 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lksx\" (UniqueName: \"kubernetes.io/projected/4c910732-ba37-4887-9ab3-32c491361007-kube-api-access-2lksx\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.849325 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c910732-ba37-4887-9ab3-32c491361007-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c910732-ba37-4887-9ab3-32c491361007\") " pod="openstack/ceilometer-0" Dec 06 08:31:45 crc kubenswrapper[4945]: I1206 08:31:45.908139 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 08:31:46 crc kubenswrapper[4945]: W1206 08:31:46.432684 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c910732_ba37_4887_9ab3_32c491361007.slice/crio-e0c55dd3a1483f09b5ffc7bdb930b4bf6f22347f3841ca7995c067dc7ae30585 WatchSource:0}: Error finding container e0c55dd3a1483f09b5ffc7bdb930b4bf6f22347f3841ca7995c067dc7ae30585: Status 404 returned error can't find the container with id e0c55dd3a1483f09b5ffc7bdb930b4bf6f22347f3841ca7995c067dc7ae30585 Dec 06 08:31:46 crc kubenswrapper[4945]: I1206 08:31:46.433646 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 08:31:46 crc kubenswrapper[4945]: I1206 08:31:46.497972 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c910732-ba37-4887-9ab3-32c491361007","Type":"ContainerStarted","Data":"e0c55dd3a1483f09b5ffc7bdb930b4bf6f22347f3841ca7995c067dc7ae30585"} Dec 06 08:31:46 crc kubenswrapper[4945]: I1206 08:31:46.970420 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76d38018-5cbb-4a61-bc47-1b773bbeb0c8" path="/var/lib/kubelet/pods/76d38018-5cbb-4a61-bc47-1b773bbeb0c8/volumes" Dec 06 08:31:47 crc kubenswrapper[4945]: I1206 08:31:47.511762 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c910732-ba37-4887-9ab3-32c491361007","Type":"ContainerStarted","Data":"8153083a7e27005f38434851ba420c50d0ded41c9caa9fe8dd0110618ec20ee9"} Dec 06 08:31:47 crc kubenswrapper[4945]: I1206 08:31:47.512072 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c910732-ba37-4887-9ab3-32c491361007","Type":"ContainerStarted","Data":"2951c9e2f67062860971542eb81d80c5aa0cae47ec15917281ad5706de59efbe"} Dec 06 08:31:48 crc kubenswrapper[4945]: I1206 08:31:48.538814 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c910732-ba37-4887-9ab3-32c491361007","Type":"ContainerStarted","Data":"30b0754f401d7a82742c377c60da70e599f409d883a81a9ff5dc1d064df5f232"} Dec 06 08:31:48 crc kubenswrapper[4945]: I1206 08:31:48.875557 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 06 08:31:49 crc kubenswrapper[4945]: I1206 08:31:49.549616 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c910732-ba37-4887-9ab3-32c491361007","Type":"ContainerStarted","Data":"0be989bb46414aed589882bff503a21ee2df559264b11ba622d046461d0ca362"} Dec 06 08:31:49 crc kubenswrapper[4945]: I1206 08:31:49.551311 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 08:31:49 crc kubenswrapper[4945]: I1206 08:31:49.586864 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.096398726 podStartE2EDuration="4.586840197s" podCreationTimestamp="2025-12-06 08:31:45 +0000 UTC" firstStartedPulling="2025-12-06 08:31:46.435119944 +0000 UTC m=+5919.889980988" lastFinishedPulling="2025-12-06 08:31:48.925561415 +0000 UTC m=+5922.380422459" observedRunningTime="2025-12-06 08:31:49.580083676 +0000 UTC m=+5923.034944740" watchObservedRunningTime="2025-12-06 08:31:49.586840197 +0000 UTC m=+5923.041701251" Dec 06 08:31:50 crc kubenswrapper[4945]: I1206 08:31:50.497703 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 06 08:31:50 crc kubenswrapper[4945]: I1206 08:31:50.796197 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 06 08:31:50 crc kubenswrapper[4945]: I1206 08:31:50.810341 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 06 08:32:08 crc kubenswrapper[4945]: I1206 08:32:08.796163 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:32:08 crc kubenswrapper[4945]: I1206 08:32:08.796819 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:32:13 crc kubenswrapper[4945]: I1206 08:32:13.044035 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-n2tgx"] Dec 06 08:32:13 crc kubenswrapper[4945]: I1206 08:32:13.053409 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-n2tgx"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.041406 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-031f-account-create-update-gxfhm"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.054715 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3cc2-account-create-update-sjmzc"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.066569 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-p7r95"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.076687 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3cc2-account-create-update-sjmzc"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.085208 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-031f-account-create-update-gxfhm"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.094500 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-a73d-account-create-update-p8rn4"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.103174 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-cbw8h"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.111568 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-p7r95"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.120753 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-a73d-account-create-update-p8rn4"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.129216 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-cbw8h"] Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.966684 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a994594-0156-406c-943d-6e97aaf35488" path="/var/lib/kubelet/pods/4a994594-0156-406c-943d-6e97aaf35488/volumes" Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.968790 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77a80ef0-e94a-4986-834d-026c7ea1bf82" path="/var/lib/kubelet/pods/77a80ef0-e94a-4986-834d-026c7ea1bf82/volumes" Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.969598 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9035d48b-9f2d-44de-b24a-55a5ab1d4c9a" path="/var/lib/kubelet/pods/9035d48b-9f2d-44de-b24a-55a5ab1d4c9a/volumes" Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.970788 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96746661-4c5d-4339-990d-56f8fa99e2ad" path="/var/lib/kubelet/pods/96746661-4c5d-4339-990d-56f8fa99e2ad/volumes" Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.972905 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3869f07-21d7-4706-9054-9bc3d2bb5818" path="/var/lib/kubelet/pods/d3869f07-21d7-4706-9054-9bc3d2bb5818/volumes" Dec 06 08:32:14 crc kubenswrapper[4945]: I1206 08:32:14.973991 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd91a488-ec3c-4918-a147-ade496239f9c" path="/var/lib/kubelet/pods/fd91a488-ec3c-4918-a147-ade496239f9c/volumes" Dec 06 08:32:15 crc kubenswrapper[4945]: I1206 08:32:15.912527 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 08:32:26 crc kubenswrapper[4945]: I1206 08:32:26.865002 4945 scope.go:117] "RemoveContainer" containerID="b8f1e7a398b593df85f262ccbf846ca72ef60342e710fa576c434fbf1c3efa37" Dec 06 08:32:26 crc kubenswrapper[4945]: I1206 08:32:26.926521 4945 scope.go:117] "RemoveContainer" containerID="59fb3a094a0c214eda54f96364742eae7f10f943d6cf12264386b024a3e8a9bf" Dec 06 08:32:26 crc kubenswrapper[4945]: I1206 08:32:26.956413 4945 scope.go:117] "RemoveContainer" containerID="9e47a6f6c052df5d7fa1ff7776516d5ba6c0ff5927dfed4d68462994c0dd93c7" Dec 06 08:32:26 crc kubenswrapper[4945]: I1206 08:32:26.998055 4945 scope.go:117] "RemoveContainer" containerID="de795a4d2fe7a5336e49879a5efe0c55825e43cb11cd494cf5e1449151e7fee1" Dec 06 08:32:27 crc kubenswrapper[4945]: I1206 08:32:27.041337 4945 scope.go:117] "RemoveContainer" containerID="b5ad938b22ccc2cf50ba0b26ca59b5f8d3b7530b9aa3147b2483c8c1a5d3e128" Dec 06 08:32:27 crc kubenswrapper[4945]: I1206 08:32:27.086438 4945 scope.go:117] "RemoveContainer" containerID="10384df17bdd621cd4c741a2622d092127bfe5be48248a0cab800ae7540058fc" Dec 06 08:32:31 crc kubenswrapper[4945]: I1206 08:32:31.071430 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ppmxf"] Dec 06 08:32:31 crc kubenswrapper[4945]: I1206 08:32:31.084795 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ppmxf"] Dec 06 08:32:32 crc kubenswrapper[4945]: I1206 08:32:32.965597 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="804fd7d1-c742-43a9-9342-f792e005248d" path="/var/lib/kubelet/pods/804fd7d1-c742-43a9-9342-f792e005248d/volumes" Dec 06 08:32:38 crc kubenswrapper[4945]: I1206 08:32:38.795747 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:32:38 crc kubenswrapper[4945]: I1206 08:32:38.796162 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.613470 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c57b4ff-7rz2x"] Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.615177 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c57b4ff-7rz2x"] Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.615247 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.637261 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.740134 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-openstack-cell1\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.740204 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-sb\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.740250 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xs8k\" (UniqueName: \"kubernetes.io/projected/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-kube-api-access-5xs8k\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.741109 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-nb\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.741366 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-config\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.741410 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-dns-svc\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.844542 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-nb\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.844608 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-config\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.844635 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-dns-svc\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.844784 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-openstack-cell1\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.844819 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-sb\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.844859 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xs8k\" (UniqueName: \"kubernetes.io/projected/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-kube-api-access-5xs8k\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.846581 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-nb\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.846618 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-dns-svc\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.847162 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-sb\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.847171 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-openstack-cell1\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.847348 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-config\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.871564 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xs8k\" (UniqueName: \"kubernetes.io/projected/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-kube-api-access-5xs8k\") pod \"dnsmasq-dns-78c57b4ff-7rz2x\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:41 crc kubenswrapper[4945]: I1206 08:32:41.956786 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:42 crc kubenswrapper[4945]: I1206 08:32:42.440938 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c57b4ff-7rz2x"] Dec 06 08:32:43 crc kubenswrapper[4945]: I1206 08:32:43.081262 4945 generic.go:334] "Generic (PLEG): container finished" podID="ca4442b9-d04a-47f8-a227-1e9fe04d53c1" containerID="a4856f5aa0d614a046a42ebd7f955afc07848d81c00d20a418f0b0d11a4bad1c" exitCode=0 Dec 06 08:32:43 crc kubenswrapper[4945]: I1206 08:32:43.081378 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" event={"ID":"ca4442b9-d04a-47f8-a227-1e9fe04d53c1","Type":"ContainerDied","Data":"a4856f5aa0d614a046a42ebd7f955afc07848d81c00d20a418f0b0d11a4bad1c"} Dec 06 08:32:43 crc kubenswrapper[4945]: I1206 08:32:43.081653 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" event={"ID":"ca4442b9-d04a-47f8-a227-1e9fe04d53c1","Type":"ContainerStarted","Data":"20586ba22e87b45fdd10a092904ba8bbf2a64db87a078fc7b841158c9f325952"} Dec 06 08:32:44 crc kubenswrapper[4945]: I1206 08:32:44.095316 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" event={"ID":"ca4442b9-d04a-47f8-a227-1e9fe04d53c1","Type":"ContainerStarted","Data":"4768f72189e8e75845b96401f1d3f3644a89991ea147ff83c0f7049bbf4101ae"} Dec 06 08:32:44 crc kubenswrapper[4945]: I1206 08:32:44.096419 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:44 crc kubenswrapper[4945]: I1206 08:32:44.122967 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" podStartSLOduration=3.122939808 podStartE2EDuration="3.122939808s" podCreationTimestamp="2025-12-06 08:32:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:32:44.111882852 +0000 UTC m=+5977.566743936" watchObservedRunningTime="2025-12-06 08:32:44.122939808 +0000 UTC m=+5977.577800882" Dec 06 08:32:49 crc kubenswrapper[4945]: I1206 08:32:49.069662 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pvdks"] Dec 06 08:32:49 crc kubenswrapper[4945]: I1206 08:32:49.086323 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pvdks"] Dec 06 08:32:50 crc kubenswrapper[4945]: I1206 08:32:50.978005 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bff2045b-07d6-45fa-93a5-578ea063487d" path="/var/lib/kubelet/pods/bff2045b-07d6-45fa-93a5-578ea063487d/volumes" Dec 06 08:32:51 crc kubenswrapper[4945]: I1206 08:32:51.959150 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.033841 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67849f8f7f-fbcsq"] Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.034116 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" podUID="91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" containerName="dnsmasq-dns" containerID="cri-o://66b1e89d0ae3f4fe048dc9563f39ffe0c4521bb95900bcf8d878792652e5fbbc" gracePeriod=10 Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.230700 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6fb75d6859-4scbt"] Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.234005 4945 generic.go:334] "Generic (PLEG): container finished" podID="91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" containerID="66b1e89d0ae3f4fe048dc9563f39ffe0c4521bb95900bcf8d878792652e5fbbc" exitCode=0 Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.235739 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" event={"ID":"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5","Type":"ContainerDied","Data":"66b1e89d0ae3f4fe048dc9563f39ffe0c4521bb95900bcf8d878792652e5fbbc"} Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.235834 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.273151 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fb75d6859-4scbt"] Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.380918 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-ovsdbserver-sb\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.381662 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-dns-svc\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.382092 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-openstack-cell1\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.382236 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-ovsdbserver-nb\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.382582 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-config\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.383050 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzmlb\" (UniqueName: \"kubernetes.io/projected/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-kube-api-access-zzmlb\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.487063 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-config\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.487338 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzmlb\" (UniqueName: \"kubernetes.io/projected/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-kube-api-access-zzmlb\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.487382 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-ovsdbserver-sb\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.487436 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-dns-svc\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.487499 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-openstack-cell1\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.487542 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-ovsdbserver-nb\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.488251 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-config\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.488388 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-ovsdbserver-sb\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.488515 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-dns-svc\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.488954 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-openstack-cell1\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.489077 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-ovsdbserver-nb\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.517877 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzmlb\" (UniqueName: \"kubernetes.io/projected/92fe4f16-f24e-437c-9f18-1fc3d4f3be78-kube-api-access-zzmlb\") pod \"dnsmasq-dns-6fb75d6859-4scbt\" (UID: \"92fe4f16-f24e-437c-9f18-1fc3d4f3be78\") " pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.609953 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.763175 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.896370 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-dns-svc\") pod \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.896587 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-sb\") pod \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.896629 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxxlf\" (UniqueName: \"kubernetes.io/projected/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-kube-api-access-rxxlf\") pod \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.896653 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-nb\") pod \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.896730 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-config\") pod \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\" (UID: \"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5\") " Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.902805 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-kube-api-access-rxxlf" (OuterVolumeSpecName: "kube-api-access-rxxlf") pod "91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" (UID: "91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5"). InnerVolumeSpecName "kube-api-access-rxxlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.957677 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" (UID: "91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.970331 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" (UID: "91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.970833 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" (UID: "91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.971744 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-config" (OuterVolumeSpecName: "config") pod "91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" (UID: "91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.999620 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.999655 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.999667 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxxlf\" (UniqueName: \"kubernetes.io/projected/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-kube-api-access-rxxlf\") on node \"crc\" DevicePath \"\"" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.999676 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:32:52 crc kubenswrapper[4945]: I1206 08:32:52.999685 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:32:53 crc kubenswrapper[4945]: I1206 08:32:53.029627 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-f59xz"] Dec 06 08:32:53 crc kubenswrapper[4945]: I1206 08:32:53.039180 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-f59xz"] Dec 06 08:32:53 crc kubenswrapper[4945]: I1206 08:32:53.166015 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fb75d6859-4scbt"] Dec 06 08:32:53 crc kubenswrapper[4945]: W1206 08:32:53.179407 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92fe4f16_f24e_437c_9f18_1fc3d4f3be78.slice/crio-8a034c492fbc1cbc73ffb74da8424514ad3b36b8dd29910ea18e1f0427d5cf7f WatchSource:0}: Error finding container 8a034c492fbc1cbc73ffb74da8424514ad3b36b8dd29910ea18e1f0427d5cf7f: Status 404 returned error can't find the container with id 8a034c492fbc1cbc73ffb74da8424514ad3b36b8dd29910ea18e1f0427d5cf7f Dec 06 08:32:53 crc kubenswrapper[4945]: I1206 08:32:53.245619 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" event={"ID":"92fe4f16-f24e-437c-9f18-1fc3d4f3be78","Type":"ContainerStarted","Data":"8a034c492fbc1cbc73ffb74da8424514ad3b36b8dd29910ea18e1f0427d5cf7f"} Dec 06 08:32:53 crc kubenswrapper[4945]: I1206 08:32:53.247789 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" event={"ID":"91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5","Type":"ContainerDied","Data":"39036e42258c5cbffa6b6954ae05fc139603ed07839d1d6acc3734d7466bf1f4"} Dec 06 08:32:53 crc kubenswrapper[4945]: I1206 08:32:53.247842 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67849f8f7f-fbcsq" Dec 06 08:32:53 crc kubenswrapper[4945]: I1206 08:32:53.247853 4945 scope.go:117] "RemoveContainer" containerID="66b1e89d0ae3f4fe048dc9563f39ffe0c4521bb95900bcf8d878792652e5fbbc" Dec 06 08:32:53 crc kubenswrapper[4945]: I1206 08:32:53.371058 4945 scope.go:117] "RemoveContainer" containerID="a6db0fbef174650b10873f487029094f8b7e172c43be62c0133bea48ffacc4c6" Dec 06 08:32:53 crc kubenswrapper[4945]: I1206 08:32:53.374030 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67849f8f7f-fbcsq"] Dec 06 08:32:53 crc kubenswrapper[4945]: I1206 08:32:53.381716 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67849f8f7f-fbcsq"] Dec 06 08:32:54 crc kubenswrapper[4945]: I1206 08:32:54.259668 4945 generic.go:334] "Generic (PLEG): container finished" podID="92fe4f16-f24e-437c-9f18-1fc3d4f3be78" containerID="57b48bab2498d6da638552590a5ed315294b2d907622a350f33f99f85bb97345" exitCode=0 Dec 06 08:32:54 crc kubenswrapper[4945]: I1206 08:32:54.259779 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" event={"ID":"92fe4f16-f24e-437c-9f18-1fc3d4f3be78","Type":"ContainerDied","Data":"57b48bab2498d6da638552590a5ed315294b2d907622a350f33f99f85bb97345"} Dec 06 08:32:54 crc kubenswrapper[4945]: I1206 08:32:54.964718 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67e6aa41-80ba-4a23-8fe9-591546bb4b04" path="/var/lib/kubelet/pods/67e6aa41-80ba-4a23-8fe9-591546bb4b04/volumes" Dec 06 08:32:54 crc kubenswrapper[4945]: I1206 08:32:54.965610 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" path="/var/lib/kubelet/pods/91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5/volumes" Dec 06 08:32:55 crc kubenswrapper[4945]: I1206 08:32:55.275682 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" event={"ID":"92fe4f16-f24e-437c-9f18-1fc3d4f3be78","Type":"ContainerStarted","Data":"9cfb7d9772da299a65c5027626fc408d48cc01c73e07b0e112074194a78eadd5"} Dec 06 08:32:55 crc kubenswrapper[4945]: I1206 08:32:55.275921 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:32:55 crc kubenswrapper[4945]: I1206 08:32:55.301555 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" podStartSLOduration=3.301534346 podStartE2EDuration="3.301534346s" podCreationTimestamp="2025-12-06 08:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:32:55.295388752 +0000 UTC m=+5988.750249816" watchObservedRunningTime="2025-12-06 08:32:55.301534346 +0000 UTC m=+5988.756395390" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.080499 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c"] Dec 06 08:32:59 crc kubenswrapper[4945]: E1206 08:32:59.081495 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" containerName="dnsmasq-dns" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.081513 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" containerName="dnsmasq-dns" Dec 06 08:32:59 crc kubenswrapper[4945]: E1206 08:32:59.081549 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" containerName="init" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.081557 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" containerName="init" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.081801 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="91ce0ceb-1f9e-4add-8de1-1cb2791bc3b5" containerName="dnsmasq-dns" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.082944 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.085790 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.085804 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.085990 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.086094 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.093948 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c"] Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.243490 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.243560 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btffn\" (UniqueName: \"kubernetes.io/projected/7f91cf82-922a-490d-861d-f1914f61b7da-kube-api-access-btffn\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.243607 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.243769 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.243835 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.346075 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.346144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btffn\" (UniqueName: \"kubernetes.io/projected/7f91cf82-922a-490d-861d-f1914f61b7da-kube-api-access-btffn\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.346183 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.346217 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.346239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.358042 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.358391 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.359860 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.362501 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.363650 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btffn\" (UniqueName: \"kubernetes.io/projected/7f91cf82-922a-490d-861d-f1914f61b7da-kube-api-access-btffn\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chr88c\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.427989 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.938406 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c"] Dec 06 08:32:59 crc kubenswrapper[4945]: W1206 08:32:59.939518 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f91cf82_922a_490d_861d_f1914f61b7da.slice/crio-b9a321d279583596f5135a19b061ee6d170859cdb45a244e2f68d3ab901cf42c WatchSource:0}: Error finding container b9a321d279583596f5135a19b061ee6d170859cdb45a244e2f68d3ab901cf42c: Status 404 returned error can't find the container with id b9a321d279583596f5135a19b061ee6d170859cdb45a244e2f68d3ab901cf42c Dec 06 08:32:59 crc kubenswrapper[4945]: I1206 08:32:59.942451 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 08:33:00 crc kubenswrapper[4945]: I1206 08:33:00.323191 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" event={"ID":"7f91cf82-922a-490d-861d-f1914f61b7da","Type":"ContainerStarted","Data":"b9a321d279583596f5135a19b061ee6d170859cdb45a244e2f68d3ab901cf42c"} Dec 06 08:33:02 crc kubenswrapper[4945]: I1206 08:33:02.612474 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6fb75d6859-4scbt" Dec 06 08:33:02 crc kubenswrapper[4945]: I1206 08:33:02.678637 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78c57b4ff-7rz2x"] Dec 06 08:33:02 crc kubenswrapper[4945]: I1206 08:33:02.678939 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" podUID="ca4442b9-d04a-47f8-a227-1e9fe04d53c1" containerName="dnsmasq-dns" containerID="cri-o://4768f72189e8e75845b96401f1d3f3644a89991ea147ff83c0f7049bbf4101ae" gracePeriod=10 Dec 06 08:33:03 crc kubenswrapper[4945]: I1206 08:33:03.354986 4945 generic.go:334] "Generic (PLEG): container finished" podID="ca4442b9-d04a-47f8-a227-1e9fe04d53c1" containerID="4768f72189e8e75845b96401f1d3f3644a89991ea147ff83c0f7049bbf4101ae" exitCode=0 Dec 06 08:33:03 crc kubenswrapper[4945]: I1206 08:33:03.355024 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" event={"ID":"ca4442b9-d04a-47f8-a227-1e9fe04d53c1","Type":"ContainerDied","Data":"4768f72189e8e75845b96401f1d3f3644a89991ea147ff83c0f7049bbf4101ae"} Dec 06 08:33:08 crc kubenswrapper[4945]: I1206 08:33:08.031707 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-kj84k"] Dec 06 08:33:08 crc kubenswrapper[4945]: I1206 08:33:08.040312 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-kj84k"] Dec 06 08:33:08 crc kubenswrapper[4945]: I1206 08:33:08.795501 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:33:08 crc kubenswrapper[4945]: I1206 08:33:08.795826 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:33:08 crc kubenswrapper[4945]: I1206 08:33:08.795870 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 08:33:08 crc kubenswrapper[4945]: I1206 08:33:08.796637 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 08:33:08 crc kubenswrapper[4945]: I1206 08:33:08.796694 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" gracePeriod=600 Dec 06 08:33:08 crc kubenswrapper[4945]: I1206 08:33:08.972033 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a133569e-416b-46a2-ab24-c1cfe9d4718d" path="/var/lib/kubelet/pods/a133569e-416b-46a2-ab24-c1cfe9d4718d/volumes" Dec 06 08:33:09 crc kubenswrapper[4945]: E1206 08:33:09.351482 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.417058 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" event={"ID":"ca4442b9-d04a-47f8-a227-1e9fe04d53c1","Type":"ContainerDied","Data":"20586ba22e87b45fdd10a092904ba8bbf2a64db87a078fc7b841158c9f325952"} Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.417377 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20586ba22e87b45fdd10a092904ba8bbf2a64db87a078fc7b841158c9f325952" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.419780 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" exitCode=0 Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.419809 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027"} Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.419833 4945 scope.go:117] "RemoveContainer" containerID="170cc98b8d9a5021f285e8b2c4ca1f3a5177f12dc09aebb616fcafb0121da204" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.420563 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:33:09 crc kubenswrapper[4945]: E1206 08:33:09.420917 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.432444 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.627019 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.735493 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-sb\") pod \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.735879 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-nb\") pod \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.736047 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-openstack-cell1\") pod \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.736089 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-config\") pod \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.736173 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-dns-svc\") pod \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.736200 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xs8k\" (UniqueName: \"kubernetes.io/projected/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-kube-api-access-5xs8k\") pod \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\" (UID: \"ca4442b9-d04a-47f8-a227-1e9fe04d53c1\") " Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.747364 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-kube-api-access-5xs8k" (OuterVolumeSpecName: "kube-api-access-5xs8k") pod "ca4442b9-d04a-47f8-a227-1e9fe04d53c1" (UID: "ca4442b9-d04a-47f8-a227-1e9fe04d53c1"). InnerVolumeSpecName "kube-api-access-5xs8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.790923 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ca4442b9-d04a-47f8-a227-1e9fe04d53c1" (UID: "ca4442b9-d04a-47f8-a227-1e9fe04d53c1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.791008 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ca4442b9-d04a-47f8-a227-1e9fe04d53c1" (UID: "ca4442b9-d04a-47f8-a227-1e9fe04d53c1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.797787 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-config" (OuterVolumeSpecName: "config") pod "ca4442b9-d04a-47f8-a227-1e9fe04d53c1" (UID: "ca4442b9-d04a-47f8-a227-1e9fe04d53c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.800494 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ca4442b9-d04a-47f8-a227-1e9fe04d53c1" (UID: "ca4442b9-d04a-47f8-a227-1e9fe04d53c1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.803847 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "ca4442b9-d04a-47f8-a227-1e9fe04d53c1" (UID: "ca4442b9-d04a-47f8-a227-1e9fe04d53c1"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.839778 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.839808 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-config\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.839819 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.839828 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xs8k\" (UniqueName: \"kubernetes.io/projected/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-kube-api-access-5xs8k\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.839838 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:09 crc kubenswrapper[4945]: I1206 08:33:09.839850 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca4442b9-d04a-47f8-a227-1e9fe04d53c1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:10 crc kubenswrapper[4945]: I1206 08:33:10.438831 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" event={"ID":"7f91cf82-922a-490d-861d-f1914f61b7da","Type":"ContainerStarted","Data":"df1cb4db77b35f52cfd1033b915609c1874fc1dbf77adb5f09d0997993d1e7f1"} Dec 06 08:33:10 crc kubenswrapper[4945]: I1206 08:33:10.446669 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" Dec 06 08:33:10 crc kubenswrapper[4945]: I1206 08:33:10.464213 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" podStartSLOduration=1.977113226 podStartE2EDuration="11.46419621s" podCreationTimestamp="2025-12-06 08:32:59 +0000 UTC" firstStartedPulling="2025-12-06 08:32:59.942176136 +0000 UTC m=+5993.397037180" lastFinishedPulling="2025-12-06 08:33:09.42925912 +0000 UTC m=+6002.884120164" observedRunningTime="2025-12-06 08:33:10.461377824 +0000 UTC m=+6003.916238868" watchObservedRunningTime="2025-12-06 08:33:10.46419621 +0000 UTC m=+6003.919057254" Dec 06 08:33:10 crc kubenswrapper[4945]: I1206 08:33:10.497349 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78c57b4ff-7rz2x"] Dec 06 08:33:10 crc kubenswrapper[4945]: I1206 08:33:10.507672 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78c57b4ff-7rz2x"] Dec 06 08:33:10 crc kubenswrapper[4945]: I1206 08:33:10.969901 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca4442b9-d04a-47f8-a227-1e9fe04d53c1" path="/var/lib/kubelet/pods/ca4442b9-d04a-47f8-a227-1e9fe04d53c1/volumes" Dec 06 08:33:11 crc kubenswrapper[4945]: I1206 08:33:11.964252 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-78c57b4ff-7rz2x" podUID="ca4442b9-d04a-47f8-a227-1e9fe04d53c1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.128:5353: i/o timeout" Dec 06 08:33:20 crc kubenswrapper[4945]: I1206 08:33:20.953797 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:33:20 crc kubenswrapper[4945]: E1206 08:33:20.954648 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:33:22 crc kubenswrapper[4945]: I1206 08:33:22.597684 4945 generic.go:334] "Generic (PLEG): container finished" podID="7f91cf82-922a-490d-861d-f1914f61b7da" containerID="df1cb4db77b35f52cfd1033b915609c1874fc1dbf77adb5f09d0997993d1e7f1" exitCode=0 Dec 06 08:33:22 crc kubenswrapper[4945]: I1206 08:33:22.597781 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" event={"ID":"7f91cf82-922a-490d-861d-f1914f61b7da","Type":"ContainerDied","Data":"df1cb4db77b35f52cfd1033b915609c1874fc1dbf77adb5f09d0997993d1e7f1"} Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.080493 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.151762 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ssh-key\") pod \"7f91cf82-922a-490d-861d-f1914f61b7da\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.151832 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btffn\" (UniqueName: \"kubernetes.io/projected/7f91cf82-922a-490d-861d-f1914f61b7da-kube-api-access-btffn\") pod \"7f91cf82-922a-490d-861d-f1914f61b7da\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.151904 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-inventory\") pod \"7f91cf82-922a-490d-861d-f1914f61b7da\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.151934 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-pre-adoption-validation-combined-ca-bundle\") pod \"7f91cf82-922a-490d-861d-f1914f61b7da\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.151990 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ceph\") pod \"7f91cf82-922a-490d-861d-f1914f61b7da\" (UID: \"7f91cf82-922a-490d-861d-f1914f61b7da\") " Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.157187 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ceph" (OuterVolumeSpecName: "ceph") pod "7f91cf82-922a-490d-861d-f1914f61b7da" (UID: "7f91cf82-922a-490d-861d-f1914f61b7da"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.157368 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f91cf82-922a-490d-861d-f1914f61b7da-kube-api-access-btffn" (OuterVolumeSpecName: "kube-api-access-btffn") pod "7f91cf82-922a-490d-861d-f1914f61b7da" (UID: "7f91cf82-922a-490d-861d-f1914f61b7da"). InnerVolumeSpecName "kube-api-access-btffn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.158535 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "7f91cf82-922a-490d-861d-f1914f61b7da" (UID: "7f91cf82-922a-490d-861d-f1914f61b7da"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.179473 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-inventory" (OuterVolumeSpecName: "inventory") pod "7f91cf82-922a-490d-861d-f1914f61b7da" (UID: "7f91cf82-922a-490d-861d-f1914f61b7da"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.181315 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7f91cf82-922a-490d-861d-f1914f61b7da" (UID: "7f91cf82-922a-490d-861d-f1914f61b7da"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.253946 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.254236 4945 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.254249 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.254258 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f91cf82-922a-490d-861d-f1914f61b7da-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.254269 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btffn\" (UniqueName: \"kubernetes.io/projected/7f91cf82-922a-490d-861d-f1914f61b7da-kube-api-access-btffn\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.618269 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" event={"ID":"7f91cf82-922a-490d-861d-f1914f61b7da","Type":"ContainerDied","Data":"b9a321d279583596f5135a19b061ee6d170859cdb45a244e2f68d3ab901cf42c"} Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.618336 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9a321d279583596f5135a19b061ee6d170859cdb45a244e2f68d3ab901cf42c" Dec 06 08:33:24 crc kubenswrapper[4945]: I1206 08:33:24.618375 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chr88c" Dec 06 08:33:27 crc kubenswrapper[4945]: I1206 08:33:27.288241 4945 scope.go:117] "RemoveContainer" containerID="d764b626224cc9f20ff1c6240e8386ae706bb83d4cef60d7c3aa322dc369c78e" Dec 06 08:33:27 crc kubenswrapper[4945]: I1206 08:33:27.351606 4945 scope.go:117] "RemoveContainer" containerID="956ce91cf576767e9b6191b62ed776e5c2f90d5bef68b0555b671358c0911ae4" Dec 06 08:33:27 crc kubenswrapper[4945]: I1206 08:33:27.399764 4945 scope.go:117] "RemoveContainer" containerID="61960e1e0cb8d6380227bae539a7086f39ca9c9894bc3ee625a151d39e45fc3e" Dec 06 08:33:27 crc kubenswrapper[4945]: I1206 08:33:27.434972 4945 scope.go:117] "RemoveContainer" containerID="7a2bd3e660d2a4969fe84d91f9fdfac941095f9f82a9feccd239f1787b60f0dc" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.119662 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq"] Dec 06 08:33:28 crc kubenswrapper[4945]: E1206 08:33:28.120216 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca4442b9-d04a-47f8-a227-1e9fe04d53c1" containerName="dnsmasq-dns" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.120238 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca4442b9-d04a-47f8-a227-1e9fe04d53c1" containerName="dnsmasq-dns" Dec 06 08:33:28 crc kubenswrapper[4945]: E1206 08:33:28.120305 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f91cf82-922a-490d-861d-f1914f61b7da" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.120316 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f91cf82-922a-490d-861d-f1914f61b7da" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 06 08:33:28 crc kubenswrapper[4945]: E1206 08:33:28.120343 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca4442b9-d04a-47f8-a227-1e9fe04d53c1" containerName="init" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.120351 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca4442b9-d04a-47f8-a227-1e9fe04d53c1" containerName="init" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.120600 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca4442b9-d04a-47f8-a227-1e9fe04d53c1" containerName="dnsmasq-dns" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.120644 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f91cf82-922a-490d-861d-f1914f61b7da" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.121608 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.125227 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.125806 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.127451 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.136609 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.140790 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq"] Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.182092 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.182148 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.182223 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7tjj\" (UniqueName: \"kubernetes.io/projected/73966546-b73a-4608-ab6c-ab094287f285-kube-api-access-g7tjj\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.182272 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.182315 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.283581 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7tjj\" (UniqueName: \"kubernetes.io/projected/73966546-b73a-4608-ab6c-ab094287f285-kube-api-access-g7tjj\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.283655 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.283678 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.283737 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.283771 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.291744 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.291840 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.292217 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.303071 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.303648 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7tjj\" (UniqueName: \"kubernetes.io/projected/73966546-b73a-4608-ab6c-ab094287f285-kube-api-access-g7tjj\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.460351 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:33:28 crc kubenswrapper[4945]: W1206 08:33:28.989217 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73966546_b73a_4608_ab6c_ab094287f285.slice/crio-04752e92cd8936e61e6b2d20ceec407b0805458214d8cc743fdc554714f57d94 WatchSource:0}: Error finding container 04752e92cd8936e61e6b2d20ceec407b0805458214d8cc743fdc554714f57d94: Status 404 returned error can't find the container with id 04752e92cd8936e61e6b2d20ceec407b0805458214d8cc743fdc554714f57d94 Dec 06 08:33:28 crc kubenswrapper[4945]: I1206 08:33:28.994932 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq"] Dec 06 08:33:29 crc kubenswrapper[4945]: I1206 08:33:29.671251 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" event={"ID":"73966546-b73a-4608-ab6c-ab094287f285","Type":"ContainerStarted","Data":"1bea89d1f51c533b57002486a633f6a98af99e323d8ef8cee5604de6824f1904"} Dec 06 08:33:29 crc kubenswrapper[4945]: I1206 08:33:29.671590 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" event={"ID":"73966546-b73a-4608-ab6c-ab094287f285","Type":"ContainerStarted","Data":"04752e92cd8936e61e6b2d20ceec407b0805458214d8cc743fdc554714f57d94"} Dec 06 08:33:29 crc kubenswrapper[4945]: I1206 08:33:29.697193 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" podStartSLOduration=1.276860039 podStartE2EDuration="1.697171754s" podCreationTimestamp="2025-12-06 08:33:28 +0000 UTC" firstStartedPulling="2025-12-06 08:33:29.000120564 +0000 UTC m=+6022.454981609" lastFinishedPulling="2025-12-06 08:33:29.42043228 +0000 UTC m=+6022.875293324" observedRunningTime="2025-12-06 08:33:29.686775906 +0000 UTC m=+6023.141636950" watchObservedRunningTime="2025-12-06 08:33:29.697171754 +0000 UTC m=+6023.152032788" Dec 06 08:33:31 crc kubenswrapper[4945]: I1206 08:33:31.955113 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:33:31 crc kubenswrapper[4945]: E1206 08:33:31.955957 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.614104 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fqfkk"] Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.618423 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.630902 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fqfkk"] Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.779695 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8h6b\" (UniqueName: \"kubernetes.io/projected/f7b70ded-2550-4534-afa9-b48ff9eefd95-kube-api-access-h8h6b\") pod \"community-operators-fqfkk\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.779778 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-catalog-content\") pod \"community-operators-fqfkk\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.779818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-utilities\") pod \"community-operators-fqfkk\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.882511 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-catalog-content\") pod \"community-operators-fqfkk\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.882625 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-utilities\") pod \"community-operators-fqfkk\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.882936 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8h6b\" (UniqueName: \"kubernetes.io/projected/f7b70ded-2550-4534-afa9-b48ff9eefd95-kube-api-access-h8h6b\") pod \"community-operators-fqfkk\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.883056 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-catalog-content\") pod \"community-operators-fqfkk\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.883139 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-utilities\") pod \"community-operators-fqfkk\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.900699 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8h6b\" (UniqueName: \"kubernetes.io/projected/f7b70ded-2550-4534-afa9-b48ff9eefd95-kube-api-access-h8h6b\") pod \"community-operators-fqfkk\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:32 crc kubenswrapper[4945]: I1206 08:33:32.941440 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:33 crc kubenswrapper[4945]: I1206 08:33:33.485010 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fqfkk"] Dec 06 08:33:33 crc kubenswrapper[4945]: I1206 08:33:33.716513 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqfkk" event={"ID":"f7b70ded-2550-4534-afa9-b48ff9eefd95","Type":"ContainerStarted","Data":"918dffe8fae9011815522a1fab8053c3275d9d366b4a998bfa05e48eda2a7f86"} Dec 06 08:33:33 crc kubenswrapper[4945]: I1206 08:33:33.716880 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqfkk" event={"ID":"f7b70ded-2550-4534-afa9-b48ff9eefd95","Type":"ContainerStarted","Data":"769effd8bb7c60c8e32fc81239f44a31fba42d86577d7c63b9d2128b54404bb0"} Dec 06 08:33:34 crc kubenswrapper[4945]: I1206 08:33:34.728811 4945 generic.go:334] "Generic (PLEG): container finished" podID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerID="918dffe8fae9011815522a1fab8053c3275d9d366b4a998bfa05e48eda2a7f86" exitCode=0 Dec 06 08:33:34 crc kubenswrapper[4945]: I1206 08:33:34.728917 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqfkk" event={"ID":"f7b70ded-2550-4534-afa9-b48ff9eefd95","Type":"ContainerDied","Data":"918dffe8fae9011815522a1fab8053c3275d9d366b4a998bfa05e48eda2a7f86"} Dec 06 08:33:35 crc kubenswrapper[4945]: I1206 08:33:35.739508 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqfkk" event={"ID":"f7b70ded-2550-4534-afa9-b48ff9eefd95","Type":"ContainerStarted","Data":"ffe7d52d5b70021c844347c3e4fe2bf99906c28c0af639980b07ba3456396b97"} Dec 06 08:33:37 crc kubenswrapper[4945]: I1206 08:33:37.759459 4945 generic.go:334] "Generic (PLEG): container finished" podID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerID="ffe7d52d5b70021c844347c3e4fe2bf99906c28c0af639980b07ba3456396b97" exitCode=0 Dec 06 08:33:37 crc kubenswrapper[4945]: I1206 08:33:37.759517 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqfkk" event={"ID":"f7b70ded-2550-4534-afa9-b48ff9eefd95","Type":"ContainerDied","Data":"ffe7d52d5b70021c844347c3e4fe2bf99906c28c0af639980b07ba3456396b97"} Dec 06 08:33:38 crc kubenswrapper[4945]: I1206 08:33:38.770881 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqfkk" event={"ID":"f7b70ded-2550-4534-afa9-b48ff9eefd95","Type":"ContainerStarted","Data":"33cbaad28fd9ad996fbd4d2cd3ab82c40a9d09dd833dafe10841ba627462425c"} Dec 06 08:33:38 crc kubenswrapper[4945]: I1206 08:33:38.804324 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fqfkk" podStartSLOduration=3.369284399 podStartE2EDuration="6.804306242s" podCreationTimestamp="2025-12-06 08:33:32 +0000 UTC" firstStartedPulling="2025-12-06 08:33:34.731742552 +0000 UTC m=+6028.186603586" lastFinishedPulling="2025-12-06 08:33:38.166764385 +0000 UTC m=+6031.621625429" observedRunningTime="2025-12-06 08:33:38.795133527 +0000 UTC m=+6032.249994571" watchObservedRunningTime="2025-12-06 08:33:38.804306242 +0000 UTC m=+6032.259167286" Dec 06 08:33:42 crc kubenswrapper[4945]: I1206 08:33:42.942477 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:42 crc kubenswrapper[4945]: I1206 08:33:42.943785 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:42 crc kubenswrapper[4945]: I1206 08:33:42.991696 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:43 crc kubenswrapper[4945]: I1206 08:33:43.956721 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:33:43 crc kubenswrapper[4945]: E1206 08:33:43.957052 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:33:43 crc kubenswrapper[4945]: I1206 08:33:43.958841 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:44 crc kubenswrapper[4945]: I1206 08:33:44.007188 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fqfkk"] Dec 06 08:33:45 crc kubenswrapper[4945]: I1206 08:33:45.835233 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fqfkk" podUID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerName="registry-server" containerID="cri-o://33cbaad28fd9ad996fbd4d2cd3ab82c40a9d09dd833dafe10841ba627462425c" gracePeriod=2 Dec 06 08:33:46 crc kubenswrapper[4945]: I1206 08:33:46.847248 4945 generic.go:334] "Generic (PLEG): container finished" podID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerID="33cbaad28fd9ad996fbd4d2cd3ab82c40a9d09dd833dafe10841ba627462425c" exitCode=0 Dec 06 08:33:46 crc kubenswrapper[4945]: I1206 08:33:46.847332 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqfkk" event={"ID":"f7b70ded-2550-4534-afa9-b48ff9eefd95","Type":"ContainerDied","Data":"33cbaad28fd9ad996fbd4d2cd3ab82c40a9d09dd833dafe10841ba627462425c"} Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.265592 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.426883 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-utilities\") pod \"f7b70ded-2550-4534-afa9-b48ff9eefd95\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.427273 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-catalog-content\") pod \"f7b70ded-2550-4534-afa9-b48ff9eefd95\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.427363 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8h6b\" (UniqueName: \"kubernetes.io/projected/f7b70ded-2550-4534-afa9-b48ff9eefd95-kube-api-access-h8h6b\") pod \"f7b70ded-2550-4534-afa9-b48ff9eefd95\" (UID: \"f7b70ded-2550-4534-afa9-b48ff9eefd95\") " Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.427909 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-utilities" (OuterVolumeSpecName: "utilities") pod "f7b70ded-2550-4534-afa9-b48ff9eefd95" (UID: "f7b70ded-2550-4534-afa9-b48ff9eefd95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.433359 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7b70ded-2550-4534-afa9-b48ff9eefd95-kube-api-access-h8h6b" (OuterVolumeSpecName: "kube-api-access-h8h6b") pod "f7b70ded-2550-4534-afa9-b48ff9eefd95" (UID: "f7b70ded-2550-4534-afa9-b48ff9eefd95"). InnerVolumeSpecName "kube-api-access-h8h6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.509893 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7b70ded-2550-4534-afa9-b48ff9eefd95" (UID: "f7b70ded-2550-4534-afa9-b48ff9eefd95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.530429 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8h6b\" (UniqueName: \"kubernetes.io/projected/f7b70ded-2550-4534-afa9-b48ff9eefd95-kube-api-access-h8h6b\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.530474 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.530486 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7b70ded-2550-4534-afa9-b48ff9eefd95-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.861930 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqfkk" event={"ID":"f7b70ded-2550-4534-afa9-b48ff9eefd95","Type":"ContainerDied","Data":"769effd8bb7c60c8e32fc81239f44a31fba42d86577d7c63b9d2128b54404bb0"} Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.862007 4945 scope.go:117] "RemoveContainer" containerID="33cbaad28fd9ad996fbd4d2cd3ab82c40a9d09dd833dafe10841ba627462425c" Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.862191 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqfkk" Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.893240 4945 scope.go:117] "RemoveContainer" containerID="ffe7d52d5b70021c844347c3e4fe2bf99906c28c0af639980b07ba3456396b97" Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.908390 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fqfkk"] Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.922192 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fqfkk"] Dec 06 08:33:47 crc kubenswrapper[4945]: I1206 08:33:47.925275 4945 scope.go:117] "RemoveContainer" containerID="918dffe8fae9011815522a1fab8053c3275d9d366b4a998bfa05e48eda2a7f86" Dec 06 08:33:48 crc kubenswrapper[4945]: I1206 08:33:48.054932 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-1115-account-create-update-qkllx"] Dec 06 08:33:48 crc kubenswrapper[4945]: I1206 08:33:48.064802 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-fmsp8"] Dec 06 08:33:48 crc kubenswrapper[4945]: I1206 08:33:48.074104 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-1115-account-create-update-qkllx"] Dec 06 08:33:48 crc kubenswrapper[4945]: I1206 08:33:48.084133 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-fmsp8"] Dec 06 08:33:48 crc kubenswrapper[4945]: I1206 08:33:48.974554 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="211109fe-4865-4082-ae88-8b4fb6facefa" path="/var/lib/kubelet/pods/211109fe-4865-4082-ae88-8b4fb6facefa/volumes" Dec 06 08:33:48 crc kubenswrapper[4945]: I1206 08:33:48.975465 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98" path="/var/lib/kubelet/pods/7c4f1b77-f8f3-490a-bc9f-f7c6829a7c98/volumes" Dec 06 08:33:48 crc kubenswrapper[4945]: I1206 08:33:48.976168 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7b70ded-2550-4534-afa9-b48ff9eefd95" path="/var/lib/kubelet/pods/f7b70ded-2550-4534-afa9-b48ff9eefd95/volumes" Dec 06 08:33:56 crc kubenswrapper[4945]: I1206 08:33:56.959765 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:33:56 crc kubenswrapper[4945]: E1206 08:33:56.960601 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:34:07 crc kubenswrapper[4945]: I1206 08:34:07.953443 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:34:07 crc kubenswrapper[4945]: E1206 08:34:07.954233 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:34:21 crc kubenswrapper[4945]: I1206 08:34:21.953537 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:34:21 crc kubenswrapper[4945]: E1206 08:34:21.954476 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:34:27 crc kubenswrapper[4945]: I1206 08:34:27.603529 4945 scope.go:117] "RemoveContainer" containerID="7f8d820a4d63eb01023bf6e8bd8dbccc9109ff256c4b4e41191428d0db651904" Dec 06 08:34:27 crc kubenswrapper[4945]: I1206 08:34:27.630933 4945 scope.go:117] "RemoveContainer" containerID="d1584e94d34ca3d0c00f1f5979f9b7a8b300da0145cea72518e598040f2e41bd" Dec 06 08:34:28 crc kubenswrapper[4945]: I1206 08:34:28.045497 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-nzzsq"] Dec 06 08:34:28 crc kubenswrapper[4945]: I1206 08:34:28.054798 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-nzzsq"] Dec 06 08:34:28 crc kubenswrapper[4945]: I1206 08:34:28.966668 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="495f6a1b-e43d-436d-88ee-0f6fe915c51e" path="/var/lib/kubelet/pods/495f6a1b-e43d-436d-88ee-0f6fe915c51e/volumes" Dec 06 08:34:36 crc kubenswrapper[4945]: I1206 08:34:36.960186 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:34:36 crc kubenswrapper[4945]: E1206 08:34:36.961044 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:34:49 crc kubenswrapper[4945]: I1206 08:34:49.953876 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:34:49 crc kubenswrapper[4945]: E1206 08:34:49.954666 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:35:04 crc kubenswrapper[4945]: I1206 08:35:04.953570 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:35:04 crc kubenswrapper[4945]: E1206 08:35:04.954509 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:35:18 crc kubenswrapper[4945]: I1206 08:35:18.954948 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:35:18 crc kubenswrapper[4945]: E1206 08:35:18.956941 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:35:27 crc kubenswrapper[4945]: I1206 08:35:27.773773 4945 scope.go:117] "RemoveContainer" containerID="3645e18c9476c9157e4e2a7cbb91bd21bda41e44efddb7ac4d59eacc6f60cc11" Dec 06 08:35:31 crc kubenswrapper[4945]: I1206 08:35:31.953437 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:35:31 crc kubenswrapper[4945]: E1206 08:35:31.954409 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:35:45 crc kubenswrapper[4945]: I1206 08:35:45.953246 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:35:45 crc kubenswrapper[4945]: E1206 08:35:45.954053 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:35:59 crc kubenswrapper[4945]: I1206 08:35:59.953923 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:35:59 crc kubenswrapper[4945]: E1206 08:35:59.954562 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:36:10 crc kubenswrapper[4945]: I1206 08:36:10.953765 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:36:10 crc kubenswrapper[4945]: E1206 08:36:10.954521 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:36:21 crc kubenswrapper[4945]: I1206 08:36:21.953159 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:36:21 crc kubenswrapper[4945]: E1206 08:36:21.955121 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.745033 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8gc5n"] Dec 06 08:36:33 crc kubenswrapper[4945]: E1206 08:36:33.746087 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerName="extract-utilities" Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.746105 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerName="extract-utilities" Dec 06 08:36:33 crc kubenswrapper[4945]: E1206 08:36:33.746125 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerName="registry-server" Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.746133 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerName="registry-server" Dec 06 08:36:33 crc kubenswrapper[4945]: E1206 08:36:33.746153 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerName="extract-content" Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.746160 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerName="extract-content" Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.746391 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7b70ded-2550-4534-afa9-b48ff9eefd95" containerName="registry-server" Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.747964 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.759704 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8gc5n"] Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.897096 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhkcd\" (UniqueName: \"kubernetes.io/projected/3c199afb-3522-4788-952a-9e6654bdb510-kube-api-access-mhkcd\") pod \"certified-operators-8gc5n\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.897249 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-utilities\") pod \"certified-operators-8gc5n\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.897387 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-catalog-content\") pod \"certified-operators-8gc5n\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:33 crc kubenswrapper[4945]: I1206 08:36:33.999776 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-utilities\") pod \"certified-operators-8gc5n\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:34 crc kubenswrapper[4945]: I1206 08:36:33.999863 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-catalog-content\") pod \"certified-operators-8gc5n\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:34 crc kubenswrapper[4945]: I1206 08:36:33.999982 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhkcd\" (UniqueName: \"kubernetes.io/projected/3c199afb-3522-4788-952a-9e6654bdb510-kube-api-access-mhkcd\") pod \"certified-operators-8gc5n\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:34 crc kubenswrapper[4945]: I1206 08:36:34.000840 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-utilities\") pod \"certified-operators-8gc5n\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:34 crc kubenswrapper[4945]: I1206 08:36:34.001079 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-catalog-content\") pod \"certified-operators-8gc5n\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:34 crc kubenswrapper[4945]: I1206 08:36:34.023211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhkcd\" (UniqueName: \"kubernetes.io/projected/3c199afb-3522-4788-952a-9e6654bdb510-kube-api-access-mhkcd\") pod \"certified-operators-8gc5n\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:34 crc kubenswrapper[4945]: I1206 08:36:34.104468 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:35 crc kubenswrapper[4945]: I1206 08:36:35.197870 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8gc5n"] Dec 06 08:36:35 crc kubenswrapper[4945]: I1206 08:36:35.548009 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c199afb-3522-4788-952a-9e6654bdb510" containerID="7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd" exitCode=0 Dec 06 08:36:35 crc kubenswrapper[4945]: I1206 08:36:35.548050 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gc5n" event={"ID":"3c199afb-3522-4788-952a-9e6654bdb510","Type":"ContainerDied","Data":"7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd"} Dec 06 08:36:35 crc kubenswrapper[4945]: I1206 08:36:35.548338 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gc5n" event={"ID":"3c199afb-3522-4788-952a-9e6654bdb510","Type":"ContainerStarted","Data":"145a1e7744bb4c162ec8e8ec1dd5de8eaccea52c3d7f06f6271e232da838ed21"} Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.566685 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gc5n" event={"ID":"3c199afb-3522-4788-952a-9e6654bdb510","Type":"ContainerStarted","Data":"c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c"} Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.750879 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w6x4q"] Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.753887 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.761752 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w6x4q"] Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.863809 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-catalog-content\") pod \"redhat-operators-w6x4q\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.864348 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf9nr\" (UniqueName: \"kubernetes.io/projected/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-kube-api-access-cf9nr\") pod \"redhat-operators-w6x4q\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.864472 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-utilities\") pod \"redhat-operators-w6x4q\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.959308 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:36:36 crc kubenswrapper[4945]: E1206 08:36:36.959599 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.966130 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf9nr\" (UniqueName: \"kubernetes.io/projected/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-kube-api-access-cf9nr\") pod \"redhat-operators-w6x4q\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.968636 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-utilities\") pod \"redhat-operators-w6x4q\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.969238 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-utilities\") pod \"redhat-operators-w6x4q\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.969534 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-catalog-content\") pod \"redhat-operators-w6x4q\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.969859 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-catalog-content\") pod \"redhat-operators-w6x4q\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:36 crc kubenswrapper[4945]: I1206 08:36:36.994292 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf9nr\" (UniqueName: \"kubernetes.io/projected/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-kube-api-access-cf9nr\") pod \"redhat-operators-w6x4q\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:37 crc kubenswrapper[4945]: I1206 08:36:37.070196 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:37 crc kubenswrapper[4945]: W1206 08:36:37.649094 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb1c753d_e9e5_4a4a_95df_4a26528a7a3e.slice/crio-b2cebe54bc466e6871216d1eb6bb7885a82f0cee97f1d4cc6f9a02012815a773 WatchSource:0}: Error finding container b2cebe54bc466e6871216d1eb6bb7885a82f0cee97f1d4cc6f9a02012815a773: Status 404 returned error can't find the container with id b2cebe54bc466e6871216d1eb6bb7885a82f0cee97f1d4cc6f9a02012815a773 Dec 06 08:36:37 crc kubenswrapper[4945]: I1206 08:36:37.653594 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w6x4q"] Dec 06 08:36:38 crc kubenswrapper[4945]: I1206 08:36:38.591995 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6x4q" event={"ID":"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e","Type":"ContainerStarted","Data":"b2cebe54bc466e6871216d1eb6bb7885a82f0cee97f1d4cc6f9a02012815a773"} Dec 06 08:36:41 crc kubenswrapper[4945]: I1206 08:36:41.637375 4945 generic.go:334] "Generic (PLEG): container finished" podID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerID="6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503" exitCode=0 Dec 06 08:36:41 crc kubenswrapper[4945]: I1206 08:36:41.637431 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6x4q" event={"ID":"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e","Type":"ContainerDied","Data":"6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503"} Dec 06 08:36:41 crc kubenswrapper[4945]: I1206 08:36:41.640878 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c199afb-3522-4788-952a-9e6654bdb510" containerID="c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c" exitCode=0 Dec 06 08:36:41 crc kubenswrapper[4945]: I1206 08:36:41.640931 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gc5n" event={"ID":"3c199afb-3522-4788-952a-9e6654bdb510","Type":"ContainerDied","Data":"c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c"} Dec 06 08:36:42 crc kubenswrapper[4945]: I1206 08:36:42.654355 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gc5n" event={"ID":"3c199afb-3522-4788-952a-9e6654bdb510","Type":"ContainerStarted","Data":"8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4"} Dec 06 08:36:42 crc kubenswrapper[4945]: I1206 08:36:42.675895 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8gc5n" podStartSLOduration=3.021258686 podStartE2EDuration="9.675876659s" podCreationTimestamp="2025-12-06 08:36:33 +0000 UTC" firstStartedPulling="2025-12-06 08:36:35.55044096 +0000 UTC m=+6209.005302004" lastFinishedPulling="2025-12-06 08:36:42.205058933 +0000 UTC m=+6215.659919977" observedRunningTime="2025-12-06 08:36:42.671887433 +0000 UTC m=+6216.126748487" watchObservedRunningTime="2025-12-06 08:36:42.675876659 +0000 UTC m=+6216.130737723" Dec 06 08:36:43 crc kubenswrapper[4945]: I1206 08:36:43.666783 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6x4q" event={"ID":"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e","Type":"ContainerStarted","Data":"a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381"} Dec 06 08:36:44 crc kubenswrapper[4945]: I1206 08:36:44.104973 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:44 crc kubenswrapper[4945]: I1206 08:36:44.105304 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:45 crc kubenswrapper[4945]: I1206 08:36:45.148523 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8gc5n" podUID="3c199afb-3522-4788-952a-9e6654bdb510" containerName="registry-server" probeResult="failure" output=< Dec 06 08:36:45 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 08:36:45 crc kubenswrapper[4945]: > Dec 06 08:36:47 crc kubenswrapper[4945]: I1206 08:36:47.707602 4945 generic.go:334] "Generic (PLEG): container finished" podID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerID="a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381" exitCode=0 Dec 06 08:36:47 crc kubenswrapper[4945]: I1206 08:36:47.707684 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6x4q" event={"ID":"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e","Type":"ContainerDied","Data":"a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381"} Dec 06 08:36:47 crc kubenswrapper[4945]: I1206 08:36:47.953219 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:36:47 crc kubenswrapper[4945]: E1206 08:36:47.953495 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:36:50 crc kubenswrapper[4945]: I1206 08:36:50.738016 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6x4q" event={"ID":"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e","Type":"ContainerStarted","Data":"255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564"} Dec 06 08:36:50 crc kubenswrapper[4945]: I1206 08:36:50.768908 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w6x4q" podStartSLOduration=6.920845344 podStartE2EDuration="14.768889576s" podCreationTimestamp="2025-12-06 08:36:36 +0000 UTC" firstStartedPulling="2025-12-06 08:36:41.640338624 +0000 UTC m=+6215.095199668" lastFinishedPulling="2025-12-06 08:36:49.488382856 +0000 UTC m=+6222.943243900" observedRunningTime="2025-12-06 08:36:50.759061263 +0000 UTC m=+6224.213922317" watchObservedRunningTime="2025-12-06 08:36:50.768889576 +0000 UTC m=+6224.223750620" Dec 06 08:36:54 crc kubenswrapper[4945]: I1206 08:36:54.167716 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:54 crc kubenswrapper[4945]: I1206 08:36:54.218984 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:36:57 crc kubenswrapper[4945]: I1206 08:36:57.070877 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:57 crc kubenswrapper[4945]: I1206 08:36:57.071264 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:36:58 crc kubenswrapper[4945]: I1206 08:36:58.120946 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w6x4q" podUID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerName="registry-server" probeResult="failure" output=< Dec 06 08:36:58 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 08:36:58 crc kubenswrapper[4945]: > Dec 06 08:36:59 crc kubenswrapper[4945]: I1206 08:36:59.738251 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8gc5n"] Dec 06 08:36:59 crc kubenswrapper[4945]: I1206 08:36:59.738542 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8gc5n" podUID="3c199afb-3522-4788-952a-9e6654bdb510" containerName="registry-server" containerID="cri-o://8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4" gracePeriod=2 Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.206206 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.210968 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-catalog-content\") pod \"3c199afb-3522-4788-952a-9e6654bdb510\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.211037 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-utilities\") pod \"3c199afb-3522-4788-952a-9e6654bdb510\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.211100 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhkcd\" (UniqueName: \"kubernetes.io/projected/3c199afb-3522-4788-952a-9e6654bdb510-kube-api-access-mhkcd\") pod \"3c199afb-3522-4788-952a-9e6654bdb510\" (UID: \"3c199afb-3522-4788-952a-9e6654bdb510\") " Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.212078 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-utilities" (OuterVolumeSpecName: "utilities") pod "3c199afb-3522-4788-952a-9e6654bdb510" (UID: "3c199afb-3522-4788-952a-9e6654bdb510"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.217367 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c199afb-3522-4788-952a-9e6654bdb510-kube-api-access-mhkcd" (OuterVolumeSpecName: "kube-api-access-mhkcd") pod "3c199afb-3522-4788-952a-9e6654bdb510" (UID: "3c199afb-3522-4788-952a-9e6654bdb510"). InnerVolumeSpecName "kube-api-access-mhkcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.259329 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c199afb-3522-4788-952a-9e6654bdb510" (UID: "3c199afb-3522-4788-952a-9e6654bdb510"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.313465 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.313894 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c199afb-3522-4788-952a-9e6654bdb510-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.313905 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhkcd\" (UniqueName: \"kubernetes.io/projected/3c199afb-3522-4788-952a-9e6654bdb510-kube-api-access-mhkcd\") on node \"crc\" DevicePath \"\"" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.836990 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c199afb-3522-4788-952a-9e6654bdb510" containerID="8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4" exitCode=0 Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.837067 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gc5n" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.837047 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gc5n" event={"ID":"3c199afb-3522-4788-952a-9e6654bdb510","Type":"ContainerDied","Data":"8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4"} Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.837409 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gc5n" event={"ID":"3c199afb-3522-4788-952a-9e6654bdb510","Type":"ContainerDied","Data":"145a1e7744bb4c162ec8e8ec1dd5de8eaccea52c3d7f06f6271e232da838ed21"} Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.837445 4945 scope.go:117] "RemoveContainer" containerID="8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.879001 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8gc5n"] Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.880885 4945 scope.go:117] "RemoveContainer" containerID="c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.888083 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8gc5n"] Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.908477 4945 scope.go:117] "RemoveContainer" containerID="7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.953214 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:37:00 crc kubenswrapper[4945]: E1206 08:37:00.953495 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.954620 4945 scope.go:117] "RemoveContainer" containerID="8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4" Dec 06 08:37:00 crc kubenswrapper[4945]: E1206 08:37:00.955196 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4\": container with ID starting with 8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4 not found: ID does not exist" containerID="8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.955225 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4"} err="failed to get container status \"8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4\": rpc error: code = NotFound desc = could not find container \"8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4\": container with ID starting with 8981fbc4fe93451234e2860388b36f1c743035433f9bd750e8d9c7a3441723f4 not found: ID does not exist" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.955246 4945 scope.go:117] "RemoveContainer" containerID="c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c" Dec 06 08:37:00 crc kubenswrapper[4945]: E1206 08:37:00.956086 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c\": container with ID starting with c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c not found: ID does not exist" containerID="c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.956116 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c"} err="failed to get container status \"c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c\": rpc error: code = NotFound desc = could not find container \"c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c\": container with ID starting with c2741b148c1c030641c1235ba15f26d9b96fbf03b3b022e1c0276cde6eb0d03c not found: ID does not exist" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.956136 4945 scope.go:117] "RemoveContainer" containerID="7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd" Dec 06 08:37:00 crc kubenswrapper[4945]: E1206 08:37:00.956734 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd\": container with ID starting with 7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd not found: ID does not exist" containerID="7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.956780 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd"} err="failed to get container status \"7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd\": rpc error: code = NotFound desc = could not find container \"7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd\": container with ID starting with 7d5bba4bec91358cb1c4f791568d079dbe6130ce86daadaa985009e33ffc48fd not found: ID does not exist" Dec 06 08:37:00 crc kubenswrapper[4945]: I1206 08:37:00.965007 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c199afb-3522-4788-952a-9e6654bdb510" path="/var/lib/kubelet/pods/3c199afb-3522-4788-952a-9e6654bdb510/volumes" Dec 06 08:37:07 crc kubenswrapper[4945]: I1206 08:37:07.121111 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:37:07 crc kubenswrapper[4945]: I1206 08:37:07.169983 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:37:08 crc kubenswrapper[4945]: I1206 08:37:08.537621 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w6x4q"] Dec 06 08:37:08 crc kubenswrapper[4945]: I1206 08:37:08.927083 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w6x4q" podUID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerName="registry-server" containerID="cri-o://255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564" gracePeriod=2 Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.430709 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.598817 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-utilities\") pod \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.599060 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf9nr\" (UniqueName: \"kubernetes.io/projected/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-kube-api-access-cf9nr\") pod \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.599123 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-catalog-content\") pod \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\" (UID: \"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e\") " Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.600265 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-utilities" (OuterVolumeSpecName: "utilities") pod "bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" (UID: "bb1c753d-e9e5-4a4a-95df-4a26528a7a3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.612551 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-kube-api-access-cf9nr" (OuterVolumeSpecName: "kube-api-access-cf9nr") pod "bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" (UID: "bb1c753d-e9e5-4a4a-95df-4a26528a7a3e"). InnerVolumeSpecName "kube-api-access-cf9nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.701417 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.701749 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf9nr\" (UniqueName: \"kubernetes.io/projected/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-kube-api-access-cf9nr\") on node \"crc\" DevicePath \"\"" Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.736843 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" (UID: "bb1c753d-e9e5-4a4a-95df-4a26528a7a3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.803729 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.941396 4945 generic.go:334] "Generic (PLEG): container finished" podID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerID="255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564" exitCode=0 Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.941444 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6x4q" event={"ID":"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e","Type":"ContainerDied","Data":"255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564"} Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.941472 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6x4q" event={"ID":"bb1c753d-e9e5-4a4a-95df-4a26528a7a3e","Type":"ContainerDied","Data":"b2cebe54bc466e6871216d1eb6bb7885a82f0cee97f1d4cc6f9a02012815a773"} Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.941493 4945 scope.go:117] "RemoveContainer" containerID="255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564" Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.941648 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w6x4q" Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.980542 4945 scope.go:117] "RemoveContainer" containerID="a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381" Dec 06 08:37:09 crc kubenswrapper[4945]: I1206 08:37:09.989469 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w6x4q"] Dec 06 08:37:10 crc kubenswrapper[4945]: I1206 08:37:10.001117 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w6x4q"] Dec 06 08:37:10 crc kubenswrapper[4945]: I1206 08:37:10.010815 4945 scope.go:117] "RemoveContainer" containerID="6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503" Dec 06 08:37:10 crc kubenswrapper[4945]: I1206 08:37:10.053059 4945 scope.go:117] "RemoveContainer" containerID="255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564" Dec 06 08:37:10 crc kubenswrapper[4945]: E1206 08:37:10.053574 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564\": container with ID starting with 255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564 not found: ID does not exist" containerID="255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564" Dec 06 08:37:10 crc kubenswrapper[4945]: I1206 08:37:10.053612 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564"} err="failed to get container status \"255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564\": rpc error: code = NotFound desc = could not find container \"255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564\": container with ID starting with 255604d3f6e45164949255de6371aaa74f567739f9e8eeda659f9b5ea153b564 not found: ID does not exist" Dec 06 08:37:10 crc kubenswrapper[4945]: I1206 08:37:10.053640 4945 scope.go:117] "RemoveContainer" containerID="a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381" Dec 06 08:37:10 crc kubenswrapper[4945]: E1206 08:37:10.053994 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381\": container with ID starting with a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381 not found: ID does not exist" containerID="a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381" Dec 06 08:37:10 crc kubenswrapper[4945]: I1206 08:37:10.054016 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381"} err="failed to get container status \"a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381\": rpc error: code = NotFound desc = could not find container \"a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381\": container with ID starting with a983c5f338497e60963f31915a8f3f058529879901890d985b64708c72d04381 not found: ID does not exist" Dec 06 08:37:10 crc kubenswrapper[4945]: I1206 08:37:10.054032 4945 scope.go:117] "RemoveContainer" containerID="6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503" Dec 06 08:37:10 crc kubenswrapper[4945]: E1206 08:37:10.054936 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503\": container with ID starting with 6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503 not found: ID does not exist" containerID="6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503" Dec 06 08:37:10 crc kubenswrapper[4945]: I1206 08:37:10.054989 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503"} err="failed to get container status \"6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503\": rpc error: code = NotFound desc = could not find container \"6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503\": container with ID starting with 6b8ed7bed7afd24fb9f23675c0ddd0a767f8cc3c0e546571e5a9b92fc8152503 not found: ID does not exist" Dec 06 08:37:10 crc kubenswrapper[4945]: I1206 08:37:10.977208 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" path="/var/lib/kubelet/pods/bb1c753d-e9e5-4a4a-95df-4a26528a7a3e/volumes" Dec 06 08:37:15 crc kubenswrapper[4945]: I1206 08:37:15.953644 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:37:15 crc kubenswrapper[4945]: E1206 08:37:15.954536 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:37:29 crc kubenswrapper[4945]: I1206 08:37:29.954143 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:37:29 crc kubenswrapper[4945]: E1206 08:37:29.955435 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:37:42 crc kubenswrapper[4945]: I1206 08:37:42.954409 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:37:42 crc kubenswrapper[4945]: E1206 08:37:42.955222 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:37:56 crc kubenswrapper[4945]: I1206 08:37:56.960782 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:37:56 crc kubenswrapper[4945]: E1206 08:37:56.961889 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:38:08 crc kubenswrapper[4945]: I1206 08:38:08.953811 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:38:10 crc kubenswrapper[4945]: I1206 08:38:10.516338 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"f24e645004712ec04355907a6cf2586a4e9b7dcf46af48834baf114e660215a2"} Dec 06 08:38:24 crc kubenswrapper[4945]: I1206 08:38:24.049580 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-nzd5c"] Dec 06 08:38:24 crc kubenswrapper[4945]: I1206 08:38:24.062218 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-7256-account-create-update-9jbzh"] Dec 06 08:38:24 crc kubenswrapper[4945]: I1206 08:38:24.072088 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-nzd5c"] Dec 06 08:38:24 crc kubenswrapper[4945]: I1206 08:38:24.081666 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-7256-account-create-update-9jbzh"] Dec 06 08:38:24 crc kubenswrapper[4945]: I1206 08:38:24.965635 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21cfae16-fc47-4e70-b4a7-42d57390890d" path="/var/lib/kubelet/pods/21cfae16-fc47-4e70-b4a7-42d57390890d/volumes" Dec 06 08:38:24 crc kubenswrapper[4945]: I1206 08:38:24.966252 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7d5cf50-be90-4e38-83f8-a42bddf21ec8" path="/var/lib/kubelet/pods/b7d5cf50-be90-4e38-83f8-a42bddf21ec8/volumes" Dec 06 08:38:27 crc kubenswrapper[4945]: I1206 08:38:27.928137 4945 scope.go:117] "RemoveContainer" containerID="1a55017d1f289cc210cd6b209f79e3b08ac403dfa99c19d1d5f30572db9c2bc0" Dec 06 08:38:27 crc kubenswrapper[4945]: I1206 08:38:27.961341 4945 scope.go:117] "RemoveContainer" containerID="d997631c28d3a68669756655cf73d65383aa2afee24a7dd42356460799b22d44" Dec 06 08:38:38 crc kubenswrapper[4945]: I1206 08:38:38.028586 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-sb9hp"] Dec 06 08:38:38 crc kubenswrapper[4945]: I1206 08:38:38.038325 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-sb9hp"] Dec 06 08:38:38 crc kubenswrapper[4945]: I1206 08:38:38.966354 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a1010c7-ceeb-47e1-b51d-a5d20bea055b" path="/var/lib/kubelet/pods/5a1010c7-ceeb-47e1-b51d-a5d20bea055b/volumes" Dec 06 08:39:28 crc kubenswrapper[4945]: I1206 08:39:28.048414 4945 scope.go:117] "RemoveContainer" containerID="4768f72189e8e75845b96401f1d3f3644a89991ea147ff83c0f7049bbf4101ae" Dec 06 08:39:28 crc kubenswrapper[4945]: I1206 08:39:28.070417 4945 scope.go:117] "RemoveContainer" containerID="a4856f5aa0d614a046a42ebd7f955afc07848d81c00d20a418f0b0d11a4bad1c" Dec 06 08:39:28 crc kubenswrapper[4945]: I1206 08:39:28.094113 4945 scope.go:117] "RemoveContainer" containerID="aa45bbe5ab5ab0e551d313b87afadbb4ad810c3ce47286c9d3904487dd18a4b6" Dec 06 08:40:38 crc kubenswrapper[4945]: I1206 08:40:38.795819 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:40:38 crc kubenswrapper[4945]: I1206 08:40:38.797349 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.088701 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-8532-account-create-update-ch9cf"] Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.098802 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-njgmq"] Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.108445 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-8532-account-create-update-ch9cf"] Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.117298 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-njgmq"] Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.408911 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-blx8z"] Dec 06 08:40:44 crc kubenswrapper[4945]: E1206 08:40:44.409667 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c199afb-3522-4788-952a-9e6654bdb510" containerName="extract-content" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.409693 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c199afb-3522-4788-952a-9e6654bdb510" containerName="extract-content" Dec 06 08:40:44 crc kubenswrapper[4945]: E1206 08:40:44.409708 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c199afb-3522-4788-952a-9e6654bdb510" containerName="extract-utilities" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.409718 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c199afb-3522-4788-952a-9e6654bdb510" containerName="extract-utilities" Dec 06 08:40:44 crc kubenswrapper[4945]: E1206 08:40:44.409740 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerName="extract-content" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.409748 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerName="extract-content" Dec 06 08:40:44 crc kubenswrapper[4945]: E1206 08:40:44.409763 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c199afb-3522-4788-952a-9e6654bdb510" containerName="registry-server" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.409770 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c199afb-3522-4788-952a-9e6654bdb510" containerName="registry-server" Dec 06 08:40:44 crc kubenswrapper[4945]: E1206 08:40:44.409808 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerName="registry-server" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.409816 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerName="registry-server" Dec 06 08:40:44 crc kubenswrapper[4945]: E1206 08:40:44.409831 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerName="extract-utilities" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.409838 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerName="extract-utilities" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.410112 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c199afb-3522-4788-952a-9e6654bdb510" containerName="registry-server" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.410140 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb1c753d-e9e5-4a4a-95df-4a26528a7a3e" containerName="registry-server" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.412078 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.422608 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-blx8z"] Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.544479 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-utilities\") pod \"redhat-marketplace-blx8z\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.545067 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-catalog-content\") pod \"redhat-marketplace-blx8z\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.545470 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngch7\" (UniqueName: \"kubernetes.io/projected/81d84cb4-b338-4330-895b-3b6bc0426fc0-kube-api-access-ngch7\") pod \"redhat-marketplace-blx8z\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.647476 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-utilities\") pod \"redhat-marketplace-blx8z\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.647640 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-catalog-content\") pod \"redhat-marketplace-blx8z\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.647792 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngch7\" (UniqueName: \"kubernetes.io/projected/81d84cb4-b338-4330-895b-3b6bc0426fc0-kube-api-access-ngch7\") pod \"redhat-marketplace-blx8z\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.648165 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-utilities\") pod \"redhat-marketplace-blx8z\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.648174 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-catalog-content\") pod \"redhat-marketplace-blx8z\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.674102 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngch7\" (UniqueName: \"kubernetes.io/projected/81d84cb4-b338-4330-895b-3b6bc0426fc0-kube-api-access-ngch7\") pod \"redhat-marketplace-blx8z\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.746714 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.967210 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14bc7530-6182-468d-9d49-ef02db905104" path="/var/lib/kubelet/pods/14bc7530-6182-468d-9d49-ef02db905104/volumes" Dec 06 08:40:44 crc kubenswrapper[4945]: I1206 08:40:44.968449 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9753090-1060-47a1-b7fd-86c494da37e7" path="/var/lib/kubelet/pods/c9753090-1060-47a1-b7fd-86c494da37e7/volumes" Dec 06 08:40:45 crc kubenswrapper[4945]: I1206 08:40:45.085430 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-blx8z"] Dec 06 08:40:46 crc kubenswrapper[4945]: I1206 08:40:46.037065 4945 generic.go:334] "Generic (PLEG): container finished" podID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerID="14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7" exitCode=0 Dec 06 08:40:46 crc kubenswrapper[4945]: I1206 08:40:46.037118 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blx8z" event={"ID":"81d84cb4-b338-4330-895b-3b6bc0426fc0","Type":"ContainerDied","Data":"14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7"} Dec 06 08:40:46 crc kubenswrapper[4945]: I1206 08:40:46.037684 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blx8z" event={"ID":"81d84cb4-b338-4330-895b-3b6bc0426fc0","Type":"ContainerStarted","Data":"82306255c439b44994b386b727f87745fa2f863eaf8e84681f5b0f32d67ab529"} Dec 06 08:40:46 crc kubenswrapper[4945]: I1206 08:40:46.040291 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 08:40:47 crc kubenswrapper[4945]: I1206 08:40:47.062208 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blx8z" event={"ID":"81d84cb4-b338-4330-895b-3b6bc0426fc0","Type":"ContainerStarted","Data":"7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471"} Dec 06 08:40:48 crc kubenswrapper[4945]: I1206 08:40:48.073303 4945 generic.go:334] "Generic (PLEG): container finished" podID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerID="7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471" exitCode=0 Dec 06 08:40:48 crc kubenswrapper[4945]: I1206 08:40:48.073349 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blx8z" event={"ID":"81d84cb4-b338-4330-895b-3b6bc0426fc0","Type":"ContainerDied","Data":"7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471"} Dec 06 08:40:49 crc kubenswrapper[4945]: I1206 08:40:49.084770 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blx8z" event={"ID":"81d84cb4-b338-4330-895b-3b6bc0426fc0","Type":"ContainerStarted","Data":"45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e"} Dec 06 08:40:49 crc kubenswrapper[4945]: I1206 08:40:49.107542 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-blx8z" podStartSLOduration=2.695914336 podStartE2EDuration="5.107527302s" podCreationTimestamp="2025-12-06 08:40:44 +0000 UTC" firstStartedPulling="2025-12-06 08:40:46.039601705 +0000 UTC m=+6459.494462749" lastFinishedPulling="2025-12-06 08:40:48.451214671 +0000 UTC m=+6461.906075715" observedRunningTime="2025-12-06 08:40:49.103236657 +0000 UTC m=+6462.558097701" watchObservedRunningTime="2025-12-06 08:40:49.107527302 +0000 UTC m=+6462.562388346" Dec 06 08:40:54 crc kubenswrapper[4945]: I1206 08:40:54.747403 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:54 crc kubenswrapper[4945]: I1206 08:40:54.747747 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:54 crc kubenswrapper[4945]: I1206 08:40:54.794759 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:55 crc kubenswrapper[4945]: I1206 08:40:55.193701 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:55 crc kubenswrapper[4945]: I1206 08:40:55.258744 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-blx8z"] Dec 06 08:40:57 crc kubenswrapper[4945]: I1206 08:40:57.044040 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-9fzp9"] Dec 06 08:40:57 crc kubenswrapper[4945]: I1206 08:40:57.052444 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-9fzp9"] Dec 06 08:40:57 crc kubenswrapper[4945]: I1206 08:40:57.157334 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-blx8z" podUID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerName="registry-server" containerID="cri-o://45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e" gracePeriod=2 Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.141030 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.169290 4945 generic.go:334] "Generic (PLEG): container finished" podID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerID="45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e" exitCode=0 Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.169333 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blx8z" event={"ID":"81d84cb4-b338-4330-895b-3b6bc0426fc0","Type":"ContainerDied","Data":"45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e"} Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.169357 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blx8z" event={"ID":"81d84cb4-b338-4330-895b-3b6bc0426fc0","Type":"ContainerDied","Data":"82306255c439b44994b386b727f87745fa2f863eaf8e84681f5b0f32d67ab529"} Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.169373 4945 scope.go:117] "RemoveContainer" containerID="45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.169635 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blx8z" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.188076 4945 scope.go:117] "RemoveContainer" containerID="7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.211295 4945 scope.go:117] "RemoveContainer" containerID="14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.244307 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-catalog-content\") pod \"81d84cb4-b338-4330-895b-3b6bc0426fc0\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.244759 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngch7\" (UniqueName: \"kubernetes.io/projected/81d84cb4-b338-4330-895b-3b6bc0426fc0-kube-api-access-ngch7\") pod \"81d84cb4-b338-4330-895b-3b6bc0426fc0\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.244886 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-utilities\") pod \"81d84cb4-b338-4330-895b-3b6bc0426fc0\" (UID: \"81d84cb4-b338-4330-895b-3b6bc0426fc0\") " Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.246095 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-utilities" (OuterVolumeSpecName: "utilities") pod "81d84cb4-b338-4330-895b-3b6bc0426fc0" (UID: "81d84cb4-b338-4330-895b-3b6bc0426fc0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.252344 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81d84cb4-b338-4330-895b-3b6bc0426fc0-kube-api-access-ngch7" (OuterVolumeSpecName: "kube-api-access-ngch7") pod "81d84cb4-b338-4330-895b-3b6bc0426fc0" (UID: "81d84cb4-b338-4330-895b-3b6bc0426fc0"). InnerVolumeSpecName "kube-api-access-ngch7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.260382 4945 scope.go:117] "RemoveContainer" containerID="45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e" Dec 06 08:40:58 crc kubenswrapper[4945]: E1206 08:40:58.260835 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e\": container with ID starting with 45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e not found: ID does not exist" containerID="45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.260907 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e"} err="failed to get container status \"45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e\": rpc error: code = NotFound desc = could not find container \"45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e\": container with ID starting with 45afaf37aaf3e3771e9dc5d8d6b3261b9896a3eb8110be66f0445cb08b9c789e not found: ID does not exist" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.260927 4945 scope.go:117] "RemoveContainer" containerID="7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471" Dec 06 08:40:58 crc kubenswrapper[4945]: E1206 08:40:58.261322 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471\": container with ID starting with 7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471 not found: ID does not exist" containerID="7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.261370 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471"} err="failed to get container status \"7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471\": rpc error: code = NotFound desc = could not find container \"7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471\": container with ID starting with 7a49b82738aff61a17075e02cf6af8671f0ad557df89fe9cc134304d7f696471 not found: ID does not exist" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.261398 4945 scope.go:117] "RemoveContainer" containerID="14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7" Dec 06 08:40:58 crc kubenswrapper[4945]: E1206 08:40:58.261714 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7\": container with ID starting with 14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7 not found: ID does not exist" containerID="14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.261737 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7"} err="failed to get container status \"14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7\": rpc error: code = NotFound desc = could not find container \"14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7\": container with ID starting with 14c53705b299b4c4fbdf980662344da7b8acec0c67b04a31ee1bf439b83de2d7 not found: ID does not exist" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.272921 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81d84cb4-b338-4330-895b-3b6bc0426fc0" (UID: "81d84cb4-b338-4330-895b-3b6bc0426fc0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.346185 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.346227 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngch7\" (UniqueName: \"kubernetes.io/projected/81d84cb4-b338-4330-895b-3b6bc0426fc0-kube-api-access-ngch7\") on node \"crc\" DevicePath \"\"" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.346241 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81d84cb4-b338-4330-895b-3b6bc0426fc0-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.504126 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-blx8z"] Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.513469 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-blx8z"] Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.967042 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81d84cb4-b338-4330-895b-3b6bc0426fc0" path="/var/lib/kubelet/pods/81d84cb4-b338-4330-895b-3b6bc0426fc0/volumes" Dec 06 08:40:58 crc kubenswrapper[4945]: I1206 08:40:58.967834 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="978dc61d-451e-49e8-9cc4-84f49906caae" path="/var/lib/kubelet/pods/978dc61d-451e-49e8-9cc4-84f49906caae/volumes" Dec 06 08:41:08 crc kubenswrapper[4945]: I1206 08:41:08.795626 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:41:08 crc kubenswrapper[4945]: I1206 08:41:08.796329 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:41:15 crc kubenswrapper[4945]: I1206 08:41:15.040143 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-73a0-account-create-update-4fjf2"] Dec 06 08:41:15 crc kubenswrapper[4945]: I1206 08:41:15.050323 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-73a0-account-create-update-4fjf2"] Dec 06 08:41:15 crc kubenswrapper[4945]: I1206 08:41:15.877912 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-t4zbd"] Dec 06 08:41:15 crc kubenswrapper[4945]: I1206 08:41:15.889152 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-t4zbd"] Dec 06 08:41:16 crc kubenswrapper[4945]: I1206 08:41:16.966220 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2170110e-abfb-4338-bc65-9b2e244f6a90" path="/var/lib/kubelet/pods/2170110e-abfb-4338-bc65-9b2e244f6a90/volumes" Dec 06 08:41:16 crc kubenswrapper[4945]: I1206 08:41:16.967324 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="defcea3b-acfe-4ea2-9080-0b70d3c567f4" path="/var/lib/kubelet/pods/defcea3b-acfe-4ea2-9080-0b70d3c567f4/volumes" Dec 06 08:41:28 crc kubenswrapper[4945]: I1206 08:41:28.035924 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-gq7rn"] Dec 06 08:41:28 crc kubenswrapper[4945]: I1206 08:41:28.044248 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-gq7rn"] Dec 06 08:41:28 crc kubenswrapper[4945]: I1206 08:41:28.222650 4945 scope.go:117] "RemoveContainer" containerID="19e5ed44e570fd355b1c520049c965a375defab39bc1f58986a6fa70cd598df2" Dec 06 08:41:28 crc kubenswrapper[4945]: I1206 08:41:28.251949 4945 scope.go:117] "RemoveContainer" containerID="37f9d2d64fb604dceb981e0071c548bfad58b7354e15b4ee927046c9521f86f6" Dec 06 08:41:28 crc kubenswrapper[4945]: I1206 08:41:28.313983 4945 scope.go:117] "RemoveContainer" containerID="a193e974e0a1534ae2f0e810dd25b74b7902c0abf069fe050356abdd47732aa8" Dec 06 08:41:28 crc kubenswrapper[4945]: I1206 08:41:28.366918 4945 scope.go:117] "RemoveContainer" containerID="493728116c44da3769a9dd432e913aba3b3693edea6960e644f640add3eaa5b2" Dec 06 08:41:28 crc kubenswrapper[4945]: I1206 08:41:28.405478 4945 scope.go:117] "RemoveContainer" containerID="28825cdd92c87d215b303d6224f8cb359bd70f1263383b50e4ab7003291b9313" Dec 06 08:41:28 crc kubenswrapper[4945]: I1206 08:41:28.967631 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81111732-b6e5-4c98-8372-58f18db7ed17" path="/var/lib/kubelet/pods/81111732-b6e5-4c98-8372-58f18db7ed17/volumes" Dec 06 08:41:38 crc kubenswrapper[4945]: I1206 08:41:38.795589 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:41:38 crc kubenswrapper[4945]: I1206 08:41:38.796137 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:41:38 crc kubenswrapper[4945]: I1206 08:41:38.796182 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 08:41:38 crc kubenswrapper[4945]: I1206 08:41:38.797065 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f24e645004712ec04355907a6cf2586a4e9b7dcf46af48834baf114e660215a2"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 08:41:38 crc kubenswrapper[4945]: I1206 08:41:38.797119 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://f24e645004712ec04355907a6cf2586a4e9b7dcf46af48834baf114e660215a2" gracePeriod=600 Dec 06 08:41:39 crc kubenswrapper[4945]: I1206 08:41:39.575075 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="f24e645004712ec04355907a6cf2586a4e9b7dcf46af48834baf114e660215a2" exitCode=0 Dec 06 08:41:39 crc kubenswrapper[4945]: I1206 08:41:39.575126 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"f24e645004712ec04355907a6cf2586a4e9b7dcf46af48834baf114e660215a2"} Dec 06 08:41:39 crc kubenswrapper[4945]: I1206 08:41:39.575165 4945 scope.go:117] "RemoveContainer" containerID="eb4d796c6f07cf6dc3cd1ab94423d6ea1dc021f09bd114deaf0c73516f09c027" Dec 06 08:41:40 crc kubenswrapper[4945]: I1206 08:41:40.595594 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876"} Dec 06 08:42:28 crc kubenswrapper[4945]: I1206 08:42:28.567659 4945 scope.go:117] "RemoveContainer" containerID="360b3ab8f0b3479da2025dd9eb9ac1d9ad5e162b412d76dfb351e5c4418e31a9" Dec 06 08:43:11 crc kubenswrapper[4945]: I1206 08:43:11.893150 4945 generic.go:334] "Generic (PLEG): container finished" podID="73966546-b73a-4608-ab6c-ab094287f285" containerID="1bea89d1f51c533b57002486a633f6a98af99e323d8ef8cee5604de6824f1904" exitCode=0 Dec 06 08:43:11 crc kubenswrapper[4945]: I1206 08:43:11.893261 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" event={"ID":"73966546-b73a-4608-ab6c-ab094287f285","Type":"ContainerDied","Data":"1bea89d1f51c533b57002486a633f6a98af99e323d8ef8cee5604de6824f1904"} Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.317266 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.352728 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7tjj\" (UniqueName: \"kubernetes.io/projected/73966546-b73a-4608-ab6c-ab094287f285-kube-api-access-g7tjj\") pod \"73966546-b73a-4608-ab6c-ab094287f285\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.352936 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ssh-key\") pod \"73966546-b73a-4608-ab6c-ab094287f285\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.352966 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-inventory\") pod \"73966546-b73a-4608-ab6c-ab094287f285\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.352985 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-tripleo-cleanup-combined-ca-bundle\") pod \"73966546-b73a-4608-ab6c-ab094287f285\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.353177 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ceph\") pod \"73966546-b73a-4608-ab6c-ab094287f285\" (UID: \"73966546-b73a-4608-ab6c-ab094287f285\") " Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.359627 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ceph" (OuterVolumeSpecName: "ceph") pod "73966546-b73a-4608-ab6c-ab094287f285" (UID: "73966546-b73a-4608-ab6c-ab094287f285"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.360529 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73966546-b73a-4608-ab6c-ab094287f285-kube-api-access-g7tjj" (OuterVolumeSpecName: "kube-api-access-g7tjj") pod "73966546-b73a-4608-ab6c-ab094287f285" (UID: "73966546-b73a-4608-ab6c-ab094287f285"). InnerVolumeSpecName "kube-api-access-g7tjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.361447 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "73966546-b73a-4608-ab6c-ab094287f285" (UID: "73966546-b73a-4608-ab6c-ab094287f285"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.388628 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-inventory" (OuterVolumeSpecName: "inventory") pod "73966546-b73a-4608-ab6c-ab094287f285" (UID: "73966546-b73a-4608-ab6c-ab094287f285"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.389215 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "73966546-b73a-4608-ab6c-ab094287f285" (UID: "73966546-b73a-4608-ab6c-ab094287f285"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.455559 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.455596 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.455611 4945 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.455627 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/73966546-b73a-4608-ab6c-ab094287f285-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.455640 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7tjj\" (UniqueName: \"kubernetes.io/projected/73966546-b73a-4608-ab6c-ab094287f285-kube-api-access-g7tjj\") on node \"crc\" DevicePath \"\"" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.912376 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" event={"ID":"73966546-b73a-4608-ab6c-ab094287f285","Type":"ContainerDied","Data":"04752e92cd8936e61e6b2d20ceec407b0805458214d8cc743fdc554714f57d94"} Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.912420 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04752e92cd8936e61e6b2d20ceec407b0805458214d8cc743fdc554714f57d94" Dec 06 08:43:13 crc kubenswrapper[4945]: I1206 08:43:13.912488 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.994552 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-pmpq5"] Dec 06 08:43:18 crc kubenswrapper[4945]: E1206 08:43:18.995399 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerName="extract-utilities" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.995413 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerName="extract-utilities" Dec 06 08:43:18 crc kubenswrapper[4945]: E1206 08:43:18.995430 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerName="extract-content" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.995436 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerName="extract-content" Dec 06 08:43:18 crc kubenswrapper[4945]: E1206 08:43:18.995464 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73966546-b73a-4608-ab6c-ab094287f285" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.995471 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="73966546-b73a-4608-ab6c-ab094287f285" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 06 08:43:18 crc kubenswrapper[4945]: E1206 08:43:18.995487 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerName="registry-server" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.995493 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerName="registry-server" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.995689 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="73966546-b73a-4608-ab6c-ab094287f285" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.995703 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="81d84cb4-b338-4330-895b-3b6bc0426fc0" containerName="registry-server" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.996400 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-pmpq5"] Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.996498 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.998731 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.998814 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.998818 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:43:18 crc kubenswrapper[4945]: I1206 08:43:18.999708 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.063717 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dh6g\" (UniqueName: \"kubernetes.io/projected/4eb91f90-eb31-42c6-b632-f4e17e175b17-kube-api-access-5dh6g\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.063899 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-inventory\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.064052 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.064122 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.064176 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ceph\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.166459 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.166555 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ceph\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.166693 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dh6g\" (UniqueName: \"kubernetes.io/projected/4eb91f90-eb31-42c6-b632-f4e17e175b17-kube-api-access-5dh6g\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.166738 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-inventory\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.166812 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.172372 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.173551 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-inventory\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.176861 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ceph\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.177013 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.184698 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dh6g\" (UniqueName: \"kubernetes.io/projected/4eb91f90-eb31-42c6-b632-f4e17e175b17-kube-api-access-5dh6g\") pod \"bootstrap-openstack-openstack-cell1-pmpq5\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.322810 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.861753 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-pmpq5"] Dec 06 08:43:19 crc kubenswrapper[4945]: I1206 08:43:19.964683 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" event={"ID":"4eb91f90-eb31-42c6-b632-f4e17e175b17","Type":"ContainerStarted","Data":"80d8ca2593071eabe83e109639f02cb019090bb77cb28dda0f115ea6c910e317"} Dec 06 08:43:20 crc kubenswrapper[4945]: I1206 08:43:20.975582 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" event={"ID":"4eb91f90-eb31-42c6-b632-f4e17e175b17","Type":"ContainerStarted","Data":"727cb3340f3eb03c0aa5a229bdd556250716f1dd22f3d80ae0451d5ffa02f15f"} Dec 06 08:43:21 crc kubenswrapper[4945]: I1206 08:43:21.002525 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" podStartSLOduration=2.566942281 podStartE2EDuration="3.002501785s" podCreationTimestamp="2025-12-06 08:43:18 +0000 UTC" firstStartedPulling="2025-12-06 08:43:19.870108776 +0000 UTC m=+6613.324969820" lastFinishedPulling="2025-12-06 08:43:20.30566828 +0000 UTC m=+6613.760529324" observedRunningTime="2025-12-06 08:43:20.99073892 +0000 UTC m=+6614.445599974" watchObservedRunningTime="2025-12-06 08:43:21.002501785 +0000 UTC m=+6614.457362829" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.401329 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xr7pt"] Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.405551 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.416015 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xr7pt"] Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.589061 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-utilities\") pod \"community-operators-xr7pt\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.589195 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhgqf\" (UniqueName: \"kubernetes.io/projected/d20402af-1fab-43a2-9484-6d3c9fd2dd28-kube-api-access-fhgqf\") pod \"community-operators-xr7pt\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.589271 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-catalog-content\") pod \"community-operators-xr7pt\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.694454 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhgqf\" (UniqueName: \"kubernetes.io/projected/d20402af-1fab-43a2-9484-6d3c9fd2dd28-kube-api-access-fhgqf\") pod \"community-operators-xr7pt\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.694553 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-catalog-content\") pod \"community-operators-xr7pt\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.694648 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-utilities\") pod \"community-operators-xr7pt\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.695232 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-catalog-content\") pod \"community-operators-xr7pt\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.697196 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-utilities\") pod \"community-operators-xr7pt\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.729083 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhgqf\" (UniqueName: \"kubernetes.io/projected/d20402af-1fab-43a2-9484-6d3c9fd2dd28-kube-api-access-fhgqf\") pod \"community-operators-xr7pt\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:00 crc kubenswrapper[4945]: I1206 08:44:00.731774 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:01 crc kubenswrapper[4945]: I1206 08:44:01.287906 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xr7pt"] Dec 06 08:44:01 crc kubenswrapper[4945]: I1206 08:44:01.593389 4945 generic.go:334] "Generic (PLEG): container finished" podID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerID="92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77" exitCode=0 Dec 06 08:44:01 crc kubenswrapper[4945]: I1206 08:44:01.593437 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xr7pt" event={"ID":"d20402af-1fab-43a2-9484-6d3c9fd2dd28","Type":"ContainerDied","Data":"92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77"} Dec 06 08:44:01 crc kubenswrapper[4945]: I1206 08:44:01.593466 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xr7pt" event={"ID":"d20402af-1fab-43a2-9484-6d3c9fd2dd28","Type":"ContainerStarted","Data":"b751f8549f7464032723bd0da2c65243509aa95ebebf6249e929866ac83f6041"} Dec 06 08:44:02 crc kubenswrapper[4945]: I1206 08:44:02.603715 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xr7pt" event={"ID":"d20402af-1fab-43a2-9484-6d3c9fd2dd28","Type":"ContainerStarted","Data":"181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9"} Dec 06 08:44:03 crc kubenswrapper[4945]: I1206 08:44:03.614742 4945 generic.go:334] "Generic (PLEG): container finished" podID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerID="181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9" exitCode=0 Dec 06 08:44:03 crc kubenswrapper[4945]: I1206 08:44:03.614786 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xr7pt" event={"ID":"d20402af-1fab-43a2-9484-6d3c9fd2dd28","Type":"ContainerDied","Data":"181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9"} Dec 06 08:44:04 crc kubenswrapper[4945]: I1206 08:44:04.626479 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xr7pt" event={"ID":"d20402af-1fab-43a2-9484-6d3c9fd2dd28","Type":"ContainerStarted","Data":"9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b"} Dec 06 08:44:04 crc kubenswrapper[4945]: I1206 08:44:04.649558 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xr7pt" podStartSLOduration=2.197178083 podStartE2EDuration="4.64953892s" podCreationTimestamp="2025-12-06 08:44:00 +0000 UTC" firstStartedPulling="2025-12-06 08:44:01.595162326 +0000 UTC m=+6655.050023360" lastFinishedPulling="2025-12-06 08:44:04.047523153 +0000 UTC m=+6657.502384197" observedRunningTime="2025-12-06 08:44:04.648350588 +0000 UTC m=+6658.103211642" watchObservedRunningTime="2025-12-06 08:44:04.64953892 +0000 UTC m=+6658.104399964" Dec 06 08:44:08 crc kubenswrapper[4945]: I1206 08:44:08.795208 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:44:08 crc kubenswrapper[4945]: I1206 08:44:08.795556 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:44:10 crc kubenswrapper[4945]: I1206 08:44:10.732640 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:10 crc kubenswrapper[4945]: I1206 08:44:10.732998 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:10 crc kubenswrapper[4945]: I1206 08:44:10.783749 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:11 crc kubenswrapper[4945]: I1206 08:44:11.751393 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:11 crc kubenswrapper[4945]: I1206 08:44:11.817393 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xr7pt"] Dec 06 08:44:13 crc kubenswrapper[4945]: I1206 08:44:13.727402 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xr7pt" podUID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerName="registry-server" containerID="cri-o://9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b" gracePeriod=2 Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.204645 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.377630 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhgqf\" (UniqueName: \"kubernetes.io/projected/d20402af-1fab-43a2-9484-6d3c9fd2dd28-kube-api-access-fhgqf\") pod \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.377695 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-utilities\") pod \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.377829 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-catalog-content\") pod \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\" (UID: \"d20402af-1fab-43a2-9484-6d3c9fd2dd28\") " Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.378570 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-utilities" (OuterVolumeSpecName: "utilities") pod "d20402af-1fab-43a2-9484-6d3c9fd2dd28" (UID: "d20402af-1fab-43a2-9484-6d3c9fd2dd28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.383732 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d20402af-1fab-43a2-9484-6d3c9fd2dd28-kube-api-access-fhgqf" (OuterVolumeSpecName: "kube-api-access-fhgqf") pod "d20402af-1fab-43a2-9484-6d3c9fd2dd28" (UID: "d20402af-1fab-43a2-9484-6d3c9fd2dd28"). InnerVolumeSpecName "kube-api-access-fhgqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.423414 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d20402af-1fab-43a2-9484-6d3c9fd2dd28" (UID: "d20402af-1fab-43a2-9484-6d3c9fd2dd28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.480402 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhgqf\" (UniqueName: \"kubernetes.io/projected/d20402af-1fab-43a2-9484-6d3c9fd2dd28-kube-api-access-fhgqf\") on node \"crc\" DevicePath \"\"" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.480432 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.480441 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20402af-1fab-43a2-9484-6d3c9fd2dd28-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.738540 4945 generic.go:334] "Generic (PLEG): container finished" podID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerID="9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b" exitCode=0 Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.738590 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xr7pt" event={"ID":"d20402af-1fab-43a2-9484-6d3c9fd2dd28","Type":"ContainerDied","Data":"9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b"} Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.738620 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xr7pt" event={"ID":"d20402af-1fab-43a2-9484-6d3c9fd2dd28","Type":"ContainerDied","Data":"b751f8549f7464032723bd0da2c65243509aa95ebebf6249e929866ac83f6041"} Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.738639 4945 scope.go:117] "RemoveContainer" containerID="9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.738798 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xr7pt" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.767489 4945 scope.go:117] "RemoveContainer" containerID="181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.774141 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xr7pt"] Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.785591 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xr7pt"] Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.793400 4945 scope.go:117] "RemoveContainer" containerID="92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.847932 4945 scope.go:117] "RemoveContainer" containerID="9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b" Dec 06 08:44:14 crc kubenswrapper[4945]: E1206 08:44:14.848530 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b\": container with ID starting with 9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b not found: ID does not exist" containerID="9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.848569 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b"} err="failed to get container status \"9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b\": rpc error: code = NotFound desc = could not find container \"9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b\": container with ID starting with 9cddbafdd9ae741aaa1ed5d2b374b4ca0094624b39e93030c09257f902202e7b not found: ID does not exist" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.848595 4945 scope.go:117] "RemoveContainer" containerID="181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9" Dec 06 08:44:14 crc kubenswrapper[4945]: E1206 08:44:14.850789 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9\": container with ID starting with 181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9 not found: ID does not exist" containerID="181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.850831 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9"} err="failed to get container status \"181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9\": rpc error: code = NotFound desc = could not find container \"181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9\": container with ID starting with 181da4e5d07c2f133558f7028b6a79afb961d7cd71cfef5676d391253d6dcfe9 not found: ID does not exist" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.850858 4945 scope.go:117] "RemoveContainer" containerID="92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77" Dec 06 08:44:14 crc kubenswrapper[4945]: E1206 08:44:14.851179 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77\": container with ID starting with 92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77 not found: ID does not exist" containerID="92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.851223 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77"} err="failed to get container status \"92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77\": rpc error: code = NotFound desc = could not find container \"92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77\": container with ID starting with 92429c2974c4ecedc6e3c7dbbed66cdac6f8a953ab100cb94c438bd2b75b8f77 not found: ID does not exist" Dec 06 08:44:14 crc kubenswrapper[4945]: I1206 08:44:14.966027 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" path="/var/lib/kubelet/pods/d20402af-1fab-43a2-9484-6d3c9fd2dd28/volumes" Dec 06 08:44:38 crc kubenswrapper[4945]: I1206 08:44:38.795759 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:44:38 crc kubenswrapper[4945]: I1206 08:44:38.796429 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.159563 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl"] Dec 06 08:45:00 crc kubenswrapper[4945]: E1206 08:45:00.160940 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerName="extract-utilities" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.160964 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerName="extract-utilities" Dec 06 08:45:00 crc kubenswrapper[4945]: E1206 08:45:00.160993 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerName="extract-content" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.161000 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerName="extract-content" Dec 06 08:45:00 crc kubenswrapper[4945]: E1206 08:45:00.161032 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerName="registry-server" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.161039 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerName="registry-server" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.161884 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20402af-1fab-43a2-9484-6d3c9fd2dd28" containerName="registry-server" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.163113 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.167745 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.172209 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.196384 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl"] Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.308753 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d680a703-dd11-479b-bd30-c2759e64d8cd-secret-volume\") pod \"collect-profiles-29416845-v4rhl\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.308861 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfhjj\" (UniqueName: \"kubernetes.io/projected/d680a703-dd11-479b-bd30-c2759e64d8cd-kube-api-access-tfhjj\") pod \"collect-profiles-29416845-v4rhl\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.309153 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d680a703-dd11-479b-bd30-c2759e64d8cd-config-volume\") pod \"collect-profiles-29416845-v4rhl\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.412595 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d680a703-dd11-479b-bd30-c2759e64d8cd-config-volume\") pod \"collect-profiles-29416845-v4rhl\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.412792 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d680a703-dd11-479b-bd30-c2759e64d8cd-secret-volume\") pod \"collect-profiles-29416845-v4rhl\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.412868 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfhjj\" (UniqueName: \"kubernetes.io/projected/d680a703-dd11-479b-bd30-c2759e64d8cd-kube-api-access-tfhjj\") pod \"collect-profiles-29416845-v4rhl\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.414589 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d680a703-dd11-479b-bd30-c2759e64d8cd-config-volume\") pod \"collect-profiles-29416845-v4rhl\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.425037 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d680a703-dd11-479b-bd30-c2759e64d8cd-secret-volume\") pod \"collect-profiles-29416845-v4rhl\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.431186 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfhjj\" (UniqueName: \"kubernetes.io/projected/d680a703-dd11-479b-bd30-c2759e64d8cd-kube-api-access-tfhjj\") pod \"collect-profiles-29416845-v4rhl\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.499225 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:00 crc kubenswrapper[4945]: I1206 08:45:00.976551 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl"] Dec 06 08:45:00 crc kubenswrapper[4945]: W1206 08:45:00.982871 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd680a703_dd11_479b_bd30_c2759e64d8cd.slice/crio-bf80f5ee7dd398773279dab7797b9a7fdfd64dca63105773698aa673225094cf WatchSource:0}: Error finding container bf80f5ee7dd398773279dab7797b9a7fdfd64dca63105773698aa673225094cf: Status 404 returned error can't find the container with id bf80f5ee7dd398773279dab7797b9a7fdfd64dca63105773698aa673225094cf Dec 06 08:45:01 crc kubenswrapper[4945]: I1206 08:45:01.166904 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" event={"ID":"d680a703-dd11-479b-bd30-c2759e64d8cd","Type":"ContainerStarted","Data":"ab876645010d526220eee00efec543a44be7649ec2c05cac429ea46f1695973e"} Dec 06 08:45:01 crc kubenswrapper[4945]: I1206 08:45:01.166955 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" event={"ID":"d680a703-dd11-479b-bd30-c2759e64d8cd","Type":"ContainerStarted","Data":"bf80f5ee7dd398773279dab7797b9a7fdfd64dca63105773698aa673225094cf"} Dec 06 08:45:01 crc kubenswrapper[4945]: I1206 08:45:01.188932 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" podStartSLOduration=1.18891602 podStartE2EDuration="1.18891602s" podCreationTimestamp="2025-12-06 08:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 08:45:01.185060327 +0000 UTC m=+6714.639921391" watchObservedRunningTime="2025-12-06 08:45:01.18891602 +0000 UTC m=+6714.643777064" Dec 06 08:45:02 crc kubenswrapper[4945]: I1206 08:45:02.177545 4945 generic.go:334] "Generic (PLEG): container finished" podID="d680a703-dd11-479b-bd30-c2759e64d8cd" containerID="ab876645010d526220eee00efec543a44be7649ec2c05cac429ea46f1695973e" exitCode=0 Dec 06 08:45:02 crc kubenswrapper[4945]: I1206 08:45:02.177651 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" event={"ID":"d680a703-dd11-479b-bd30-c2759e64d8cd","Type":"ContainerDied","Data":"ab876645010d526220eee00efec543a44be7649ec2c05cac429ea46f1695973e"} Dec 06 08:45:03 crc kubenswrapper[4945]: I1206 08:45:03.492378 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:03 crc kubenswrapper[4945]: I1206 08:45:03.683522 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfhjj\" (UniqueName: \"kubernetes.io/projected/d680a703-dd11-479b-bd30-c2759e64d8cd-kube-api-access-tfhjj\") pod \"d680a703-dd11-479b-bd30-c2759e64d8cd\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " Dec 06 08:45:03 crc kubenswrapper[4945]: I1206 08:45:03.684769 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d680a703-dd11-479b-bd30-c2759e64d8cd-config-volume\") pod \"d680a703-dd11-479b-bd30-c2759e64d8cd\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " Dec 06 08:45:03 crc kubenswrapper[4945]: I1206 08:45:03.684902 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d680a703-dd11-479b-bd30-c2759e64d8cd-secret-volume\") pod \"d680a703-dd11-479b-bd30-c2759e64d8cd\" (UID: \"d680a703-dd11-479b-bd30-c2759e64d8cd\") " Dec 06 08:45:03 crc kubenswrapper[4945]: I1206 08:45:03.685549 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d680a703-dd11-479b-bd30-c2759e64d8cd-config-volume" (OuterVolumeSpecName: "config-volume") pod "d680a703-dd11-479b-bd30-c2759e64d8cd" (UID: "d680a703-dd11-479b-bd30-c2759e64d8cd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:45:03 crc kubenswrapper[4945]: I1206 08:45:03.690170 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d680a703-dd11-479b-bd30-c2759e64d8cd-kube-api-access-tfhjj" (OuterVolumeSpecName: "kube-api-access-tfhjj") pod "d680a703-dd11-479b-bd30-c2759e64d8cd" (UID: "d680a703-dd11-479b-bd30-c2759e64d8cd"). InnerVolumeSpecName "kube-api-access-tfhjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:45:03 crc kubenswrapper[4945]: I1206 08:45:03.699459 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d680a703-dd11-479b-bd30-c2759e64d8cd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d680a703-dd11-479b-bd30-c2759e64d8cd" (UID: "d680a703-dd11-479b-bd30-c2759e64d8cd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:45:03 crc kubenswrapper[4945]: I1206 08:45:03.787053 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfhjj\" (UniqueName: \"kubernetes.io/projected/d680a703-dd11-479b-bd30-c2759e64d8cd-kube-api-access-tfhjj\") on node \"crc\" DevicePath \"\"" Dec 06 08:45:03 crc kubenswrapper[4945]: I1206 08:45:03.787095 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d680a703-dd11-479b-bd30-c2759e64d8cd-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 08:45:03 crc kubenswrapper[4945]: I1206 08:45:03.787105 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d680a703-dd11-479b-bd30-c2759e64d8cd-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 08:45:04 crc kubenswrapper[4945]: I1206 08:45:04.200367 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" event={"ID":"d680a703-dd11-479b-bd30-c2759e64d8cd","Type":"ContainerDied","Data":"bf80f5ee7dd398773279dab7797b9a7fdfd64dca63105773698aa673225094cf"} Dec 06 08:45:04 crc kubenswrapper[4945]: I1206 08:45:04.200415 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf80f5ee7dd398773279dab7797b9a7fdfd64dca63105773698aa673225094cf" Dec 06 08:45:04 crc kubenswrapper[4945]: I1206 08:45:04.200480 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl" Dec 06 08:45:04 crc kubenswrapper[4945]: I1206 08:45:04.264652 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng"] Dec 06 08:45:04 crc kubenswrapper[4945]: I1206 08:45:04.272870 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416800-d9jng"] Dec 06 08:45:04 crc kubenswrapper[4945]: I1206 08:45:04.966452 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78f0df6a-316c-4970-b612-3adb2e6cfb45" path="/var/lib/kubelet/pods/78f0df6a-316c-4970-b612-3adb2e6cfb45/volumes" Dec 06 08:45:08 crc kubenswrapper[4945]: I1206 08:45:08.796120 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:45:08 crc kubenswrapper[4945]: I1206 08:45:08.796698 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:45:08 crc kubenswrapper[4945]: I1206 08:45:08.796744 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 08:45:08 crc kubenswrapper[4945]: I1206 08:45:08.797493 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 08:45:08 crc kubenswrapper[4945]: I1206 08:45:08.797558 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" gracePeriod=600 Dec 06 08:45:08 crc kubenswrapper[4945]: E1206 08:45:08.927457 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:45:09 crc kubenswrapper[4945]: I1206 08:45:09.247371 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" exitCode=0 Dec 06 08:45:09 crc kubenswrapper[4945]: I1206 08:45:09.247420 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876"} Dec 06 08:45:09 crc kubenswrapper[4945]: I1206 08:45:09.247456 4945 scope.go:117] "RemoveContainer" containerID="f24e645004712ec04355907a6cf2586a4e9b7dcf46af48834baf114e660215a2" Dec 06 08:45:09 crc kubenswrapper[4945]: I1206 08:45:09.248419 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:45:09 crc kubenswrapper[4945]: E1206 08:45:09.250660 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:45:23 crc kubenswrapper[4945]: I1206 08:45:23.953074 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:45:23 crc kubenswrapper[4945]: E1206 08:45:23.953976 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:45:28 crc kubenswrapper[4945]: I1206 08:45:28.725170 4945 scope.go:117] "RemoveContainer" containerID="0f4d8a56020be620bcf286ab050f74813cdeea9e235df4abb0fcb795d5debe8c" Dec 06 08:45:38 crc kubenswrapper[4945]: I1206 08:45:38.954126 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:45:38 crc kubenswrapper[4945]: E1206 08:45:38.954927 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:45:51 crc kubenswrapper[4945]: I1206 08:45:51.954138 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:45:51 crc kubenswrapper[4945]: E1206 08:45:51.954985 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:46:03 crc kubenswrapper[4945]: I1206 08:46:03.953620 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:46:03 crc kubenswrapper[4945]: E1206 08:46:03.954368 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:46:16 crc kubenswrapper[4945]: I1206 08:46:16.971593 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:46:16 crc kubenswrapper[4945]: E1206 08:46:16.978948 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:46:25 crc kubenswrapper[4945]: I1206 08:46:25.926413 4945 generic.go:334] "Generic (PLEG): container finished" podID="4eb91f90-eb31-42c6-b632-f4e17e175b17" containerID="727cb3340f3eb03c0aa5a229bdd556250716f1dd22f3d80ae0451d5ffa02f15f" exitCode=0 Dec 06 08:46:25 crc kubenswrapper[4945]: I1206 08:46:25.926526 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" event={"ID":"4eb91f90-eb31-42c6-b632-f4e17e175b17","Type":"ContainerDied","Data":"727cb3340f3eb03c0aa5a229bdd556250716f1dd22f3d80ae0451d5ffa02f15f"} Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.540211 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.697498 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ceph\") pod \"4eb91f90-eb31-42c6-b632-f4e17e175b17\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.697609 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-inventory\") pod \"4eb91f90-eb31-42c6-b632-f4e17e175b17\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.697703 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-bootstrap-combined-ca-bundle\") pod \"4eb91f90-eb31-42c6-b632-f4e17e175b17\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.697865 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ssh-key\") pod \"4eb91f90-eb31-42c6-b632-f4e17e175b17\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.697940 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dh6g\" (UniqueName: \"kubernetes.io/projected/4eb91f90-eb31-42c6-b632-f4e17e175b17-kube-api-access-5dh6g\") pod \"4eb91f90-eb31-42c6-b632-f4e17e175b17\" (UID: \"4eb91f90-eb31-42c6-b632-f4e17e175b17\") " Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.703483 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4eb91f90-eb31-42c6-b632-f4e17e175b17" (UID: "4eb91f90-eb31-42c6-b632-f4e17e175b17"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.705710 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ceph" (OuterVolumeSpecName: "ceph") pod "4eb91f90-eb31-42c6-b632-f4e17e175b17" (UID: "4eb91f90-eb31-42c6-b632-f4e17e175b17"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.715726 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eb91f90-eb31-42c6-b632-f4e17e175b17-kube-api-access-5dh6g" (OuterVolumeSpecName: "kube-api-access-5dh6g") pod "4eb91f90-eb31-42c6-b632-f4e17e175b17" (UID: "4eb91f90-eb31-42c6-b632-f4e17e175b17"). InnerVolumeSpecName "kube-api-access-5dh6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.740131 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-inventory" (OuterVolumeSpecName: "inventory") pod "4eb91f90-eb31-42c6-b632-f4e17e175b17" (UID: "4eb91f90-eb31-42c6-b632-f4e17e175b17"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.745649 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4eb91f90-eb31-42c6-b632-f4e17e175b17" (UID: "4eb91f90-eb31-42c6-b632-f4e17e175b17"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.800358 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.800393 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dh6g\" (UniqueName: \"kubernetes.io/projected/4eb91f90-eb31-42c6-b632-f4e17e175b17-kube-api-access-5dh6g\") on node \"crc\" DevicePath \"\"" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.800409 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.800420 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.800431 4945 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb91f90-eb31-42c6-b632-f4e17e175b17-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.947648 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" event={"ID":"4eb91f90-eb31-42c6-b632-f4e17e175b17","Type":"ContainerDied","Data":"80d8ca2593071eabe83e109639f02cb019090bb77cb28dda0f115ea6c910e317"} Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.947687 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80d8ca2593071eabe83e109639f02cb019090bb77cb28dda0f115ea6c910e317" Dec 06 08:46:27 crc kubenswrapper[4945]: I1206 08:46:27.947731 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-pmpq5" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.023135 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-hhcvg"] Dec 06 08:46:28 crc kubenswrapper[4945]: E1206 08:46:28.023562 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb91f90-eb31-42c6-b632-f4e17e175b17" containerName="bootstrap-openstack-openstack-cell1" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.023578 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb91f90-eb31-42c6-b632-f4e17e175b17" containerName="bootstrap-openstack-openstack-cell1" Dec 06 08:46:28 crc kubenswrapper[4945]: E1206 08:46:28.023587 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d680a703-dd11-479b-bd30-c2759e64d8cd" containerName="collect-profiles" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.023594 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d680a703-dd11-479b-bd30-c2759e64d8cd" containerName="collect-profiles" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.023785 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eb91f90-eb31-42c6-b632-f4e17e175b17" containerName="bootstrap-openstack-openstack-cell1" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.023806 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d680a703-dd11-479b-bd30-c2759e64d8cd" containerName="collect-profiles" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.024489 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.026170 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.026593 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.026736 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.028493 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.033714 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-hhcvg"] Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.106069 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-inventory\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.106357 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ceph\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.106389 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85jw8\" (UniqueName: \"kubernetes.io/projected/7775a907-216c-477c-9087-4cd2393ee743-kube-api-access-85jw8\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.106477 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ssh-key\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.208720 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ssh-key\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.209082 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-inventory\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.209207 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ceph\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.209236 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85jw8\" (UniqueName: \"kubernetes.io/projected/7775a907-216c-477c-9087-4cd2393ee743-kube-api-access-85jw8\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.213152 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ceph\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.216053 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ssh-key\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.219292 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-inventory\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.234213 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85jw8\" (UniqueName: \"kubernetes.io/projected/7775a907-216c-477c-9087-4cd2393ee743-kube-api-access-85jw8\") pod \"download-cache-openstack-openstack-cell1-hhcvg\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.344841 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.941468 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-hhcvg"] Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.956386 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:46:28 crc kubenswrapper[4945]: E1206 08:46:28.956640 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:46:28 crc kubenswrapper[4945]: I1206 08:46:28.957495 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 08:46:29 crc kubenswrapper[4945]: I1206 08:46:29.972015 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" event={"ID":"7775a907-216c-477c-9087-4cd2393ee743","Type":"ContainerStarted","Data":"c21dd667919933f7d647dc863e896bc992a9b94e520b9beec4fcb1f890fd2bea"} Dec 06 08:46:29 crc kubenswrapper[4945]: I1206 08:46:29.973204 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" event={"ID":"7775a907-216c-477c-9087-4cd2393ee743","Type":"ContainerStarted","Data":"e6c7c4b04e2d8683cdc68e10009afd896b575729885064a14d2bc7f5dcaa0508"} Dec 06 08:46:29 crc kubenswrapper[4945]: I1206 08:46:29.996709 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" podStartSLOduration=1.5214773529999999 podStartE2EDuration="1.996688548s" podCreationTimestamp="2025-12-06 08:46:28 +0000 UTC" firstStartedPulling="2025-12-06 08:46:28.957208055 +0000 UTC m=+6802.412069099" lastFinishedPulling="2025-12-06 08:46:29.43241925 +0000 UTC m=+6802.887280294" observedRunningTime="2025-12-06 08:46:29.989949768 +0000 UTC m=+6803.444810822" watchObservedRunningTime="2025-12-06 08:46:29.996688548 +0000 UTC m=+6803.451549602" Dec 06 08:46:39 crc kubenswrapper[4945]: I1206 08:46:39.953263 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:46:39 crc kubenswrapper[4945]: E1206 08:46:39.954168 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.220661 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zc782"] Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.223936 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.230368 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zc782"] Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.293080 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-utilities\") pod \"redhat-operators-zc782\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.296590 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qklnl\" (UniqueName: \"kubernetes.io/projected/d040649e-1c02-4d63-9f68-861cf49a7013-kube-api-access-qklnl\") pod \"redhat-operators-zc782\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.296703 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-catalog-content\") pod \"redhat-operators-zc782\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.399373 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-utilities\") pod \"redhat-operators-zc782\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.399445 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qklnl\" (UniqueName: \"kubernetes.io/projected/d040649e-1c02-4d63-9f68-861cf49a7013-kube-api-access-qklnl\") pod \"redhat-operators-zc782\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.399464 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-catalog-content\") pod \"redhat-operators-zc782\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.399992 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-utilities\") pod \"redhat-operators-zc782\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.400026 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-catalog-content\") pod \"redhat-operators-zc782\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.420843 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qklnl\" (UniqueName: \"kubernetes.io/projected/d040649e-1c02-4d63-9f68-861cf49a7013-kube-api-access-qklnl\") pod \"redhat-operators-zc782\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:52 crc kubenswrapper[4945]: I1206 08:46:52.580016 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:46:53 crc kubenswrapper[4945]: I1206 08:46:53.131252 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zc782"] Dec 06 08:46:53 crc kubenswrapper[4945]: I1206 08:46:53.190998 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc782" event={"ID":"d040649e-1c02-4d63-9f68-861cf49a7013","Type":"ContainerStarted","Data":"a1fabe3c18694455ed64f8820bc616c8c00e84a8d06702af2de6e70580ceb197"} Dec 06 08:46:53 crc kubenswrapper[4945]: I1206 08:46:53.953614 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:46:53 crc kubenswrapper[4945]: E1206 08:46:53.954251 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:46:54 crc kubenswrapper[4945]: I1206 08:46:54.202421 4945 generic.go:334] "Generic (PLEG): container finished" podID="d040649e-1c02-4d63-9f68-861cf49a7013" containerID="cddb51db5fb87d95d60ec223f9dce215815f99283edbbea586ed5d95ab5484c0" exitCode=0 Dec 06 08:46:54 crc kubenswrapper[4945]: I1206 08:46:54.202463 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc782" event={"ID":"d040649e-1c02-4d63-9f68-861cf49a7013","Type":"ContainerDied","Data":"cddb51db5fb87d95d60ec223f9dce215815f99283edbbea586ed5d95ab5484c0"} Dec 06 08:46:55 crc kubenswrapper[4945]: I1206 08:46:55.213893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc782" event={"ID":"d040649e-1c02-4d63-9f68-861cf49a7013","Type":"ContainerStarted","Data":"534cacad4e5667650370d503496f54e9c5adea2dcdddc829d2bce394d9fa67ed"} Dec 06 08:46:59 crc kubenswrapper[4945]: I1206 08:46:59.248857 4945 generic.go:334] "Generic (PLEG): container finished" podID="d040649e-1c02-4d63-9f68-861cf49a7013" containerID="534cacad4e5667650370d503496f54e9c5adea2dcdddc829d2bce394d9fa67ed" exitCode=0 Dec 06 08:46:59 crc kubenswrapper[4945]: I1206 08:46:59.248961 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc782" event={"ID":"d040649e-1c02-4d63-9f68-861cf49a7013","Type":"ContainerDied","Data":"534cacad4e5667650370d503496f54e9c5adea2dcdddc829d2bce394d9fa67ed"} Dec 06 08:47:01 crc kubenswrapper[4945]: I1206 08:47:01.276748 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc782" event={"ID":"d040649e-1c02-4d63-9f68-861cf49a7013","Type":"ContainerStarted","Data":"4141ed262ec41a6c4b5b0d9c387db19a5c11df42193ff360c044f2a9c1ac0a37"} Dec 06 08:47:01 crc kubenswrapper[4945]: I1206 08:47:01.302167 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zc782" podStartSLOduration=3.354414033 podStartE2EDuration="9.302149844s" podCreationTimestamp="2025-12-06 08:46:52 +0000 UTC" firstStartedPulling="2025-12-06 08:46:54.204937937 +0000 UTC m=+6827.659798971" lastFinishedPulling="2025-12-06 08:47:00.152673738 +0000 UTC m=+6833.607534782" observedRunningTime="2025-12-06 08:47:01.297668874 +0000 UTC m=+6834.752529928" watchObservedRunningTime="2025-12-06 08:47:01.302149844 +0000 UTC m=+6834.757010888" Dec 06 08:47:02 crc kubenswrapper[4945]: I1206 08:47:02.580401 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:47:02 crc kubenswrapper[4945]: I1206 08:47:02.580459 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:47:03 crc kubenswrapper[4945]: I1206 08:47:03.631705 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zc782" podUID="d040649e-1c02-4d63-9f68-861cf49a7013" containerName="registry-server" probeResult="failure" output=< Dec 06 08:47:03 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 08:47:03 crc kubenswrapper[4945]: > Dec 06 08:47:05 crc kubenswrapper[4945]: I1206 08:47:05.953300 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:47:05 crc kubenswrapper[4945]: E1206 08:47:05.953918 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:47:12 crc kubenswrapper[4945]: I1206 08:47:12.635021 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:47:12 crc kubenswrapper[4945]: I1206 08:47:12.691306 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:47:12 crc kubenswrapper[4945]: I1206 08:47:12.879791 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zc782"] Dec 06 08:47:14 crc kubenswrapper[4945]: I1206 08:47:14.408739 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zc782" podUID="d040649e-1c02-4d63-9f68-861cf49a7013" containerName="registry-server" containerID="cri-o://4141ed262ec41a6c4b5b0d9c387db19a5c11df42193ff360c044f2a9c1ac0a37" gracePeriod=2 Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.435412 4945 generic.go:334] "Generic (PLEG): container finished" podID="d040649e-1c02-4d63-9f68-861cf49a7013" containerID="4141ed262ec41a6c4b5b0d9c387db19a5c11df42193ff360c044f2a9c1ac0a37" exitCode=0 Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.435489 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc782" event={"ID":"d040649e-1c02-4d63-9f68-861cf49a7013","Type":"ContainerDied","Data":"4141ed262ec41a6c4b5b0d9c387db19a5c11df42193ff360c044f2a9c1ac0a37"} Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.719202 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.746385 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-catalog-content\") pod \"d040649e-1c02-4d63-9f68-861cf49a7013\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.746428 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-utilities\") pod \"d040649e-1c02-4d63-9f68-861cf49a7013\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.746487 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qklnl\" (UniqueName: \"kubernetes.io/projected/d040649e-1c02-4d63-9f68-861cf49a7013-kube-api-access-qklnl\") pod \"d040649e-1c02-4d63-9f68-861cf49a7013\" (UID: \"d040649e-1c02-4d63-9f68-861cf49a7013\") " Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.748145 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-utilities" (OuterVolumeSpecName: "utilities") pod "d040649e-1c02-4d63-9f68-861cf49a7013" (UID: "d040649e-1c02-4d63-9f68-861cf49a7013"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.755872 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d040649e-1c02-4d63-9f68-861cf49a7013-kube-api-access-qklnl" (OuterVolumeSpecName: "kube-api-access-qklnl") pod "d040649e-1c02-4d63-9f68-861cf49a7013" (UID: "d040649e-1c02-4d63-9f68-861cf49a7013"). InnerVolumeSpecName "kube-api-access-qklnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.849490 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.849541 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qklnl\" (UniqueName: \"kubernetes.io/projected/d040649e-1c02-4d63-9f68-861cf49a7013-kube-api-access-qklnl\") on node \"crc\" DevicePath \"\"" Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.860370 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d040649e-1c02-4d63-9f68-861cf49a7013" (UID: "d040649e-1c02-4d63-9f68-861cf49a7013"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:47:17 crc kubenswrapper[4945]: I1206 08:47:17.951179 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d040649e-1c02-4d63-9f68-861cf49a7013-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:47:18 crc kubenswrapper[4945]: I1206 08:47:18.447370 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc782" event={"ID":"d040649e-1c02-4d63-9f68-861cf49a7013","Type":"ContainerDied","Data":"a1fabe3c18694455ed64f8820bc616c8c00e84a8d06702af2de6e70580ceb197"} Dec 06 08:47:18 crc kubenswrapper[4945]: I1206 08:47:18.447429 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc782" Dec 06 08:47:18 crc kubenswrapper[4945]: I1206 08:47:18.448395 4945 scope.go:117] "RemoveContainer" containerID="4141ed262ec41a6c4b5b0d9c387db19a5c11df42193ff360c044f2a9c1ac0a37" Dec 06 08:47:18 crc kubenswrapper[4945]: I1206 08:47:18.470229 4945 scope.go:117] "RemoveContainer" containerID="534cacad4e5667650370d503496f54e9c5adea2dcdddc829d2bce394d9fa67ed" Dec 06 08:47:18 crc kubenswrapper[4945]: I1206 08:47:18.486629 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zc782"] Dec 06 08:47:18 crc kubenswrapper[4945]: I1206 08:47:18.494522 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zc782"] Dec 06 08:47:18 crc kubenswrapper[4945]: I1206 08:47:18.529884 4945 scope.go:117] "RemoveContainer" containerID="cddb51db5fb87d95d60ec223f9dce215815f99283edbbea586ed5d95ab5484c0" Dec 06 08:47:18 crc kubenswrapper[4945]: I1206 08:47:18.954203 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:47:18 crc kubenswrapper[4945]: E1206 08:47:18.954915 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:47:18 crc kubenswrapper[4945]: I1206 08:47:18.966767 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d040649e-1c02-4d63-9f68-861cf49a7013" path="/var/lib/kubelet/pods/d040649e-1c02-4d63-9f68-861cf49a7013/volumes" Dec 06 08:47:29 crc kubenswrapper[4945]: I1206 08:47:29.953958 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:47:29 crc kubenswrapper[4945]: E1206 08:47:29.954938 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:47:41 crc kubenswrapper[4945]: I1206 08:47:41.952932 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:47:41 crc kubenswrapper[4945]: E1206 08:47:41.953781 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:47:53 crc kubenswrapper[4945]: I1206 08:47:53.953909 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:47:53 crc kubenswrapper[4945]: E1206 08:47:53.954718 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:48:00 crc kubenswrapper[4945]: I1206 08:48:00.857492 4945 generic.go:334] "Generic (PLEG): container finished" podID="7775a907-216c-477c-9087-4cd2393ee743" containerID="c21dd667919933f7d647dc863e896bc992a9b94e520b9beec4fcb1f890fd2bea" exitCode=0 Dec 06 08:48:00 crc kubenswrapper[4945]: I1206 08:48:00.857578 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" event={"ID":"7775a907-216c-477c-9087-4cd2393ee743","Type":"ContainerDied","Data":"c21dd667919933f7d647dc863e896bc992a9b94e520b9beec4fcb1f890fd2bea"} Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.304332 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.459933 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-inventory\") pod \"7775a907-216c-477c-9087-4cd2393ee743\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.460063 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ssh-key\") pod \"7775a907-216c-477c-9087-4cd2393ee743\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.460126 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85jw8\" (UniqueName: \"kubernetes.io/projected/7775a907-216c-477c-9087-4cd2393ee743-kube-api-access-85jw8\") pod \"7775a907-216c-477c-9087-4cd2393ee743\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.460235 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ceph\") pod \"7775a907-216c-477c-9087-4cd2393ee743\" (UID: \"7775a907-216c-477c-9087-4cd2393ee743\") " Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.466618 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7775a907-216c-477c-9087-4cd2393ee743-kube-api-access-85jw8" (OuterVolumeSpecName: "kube-api-access-85jw8") pod "7775a907-216c-477c-9087-4cd2393ee743" (UID: "7775a907-216c-477c-9087-4cd2393ee743"). InnerVolumeSpecName "kube-api-access-85jw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.466821 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ceph" (OuterVolumeSpecName: "ceph") pod "7775a907-216c-477c-9087-4cd2393ee743" (UID: "7775a907-216c-477c-9087-4cd2393ee743"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.486978 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7775a907-216c-477c-9087-4cd2393ee743" (UID: "7775a907-216c-477c-9087-4cd2393ee743"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.488917 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-inventory" (OuterVolumeSpecName: "inventory") pod "7775a907-216c-477c-9087-4cd2393ee743" (UID: "7775a907-216c-477c-9087-4cd2393ee743"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.562949 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.563172 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.563240 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85jw8\" (UniqueName: \"kubernetes.io/projected/7775a907-216c-477c-9087-4cd2393ee743-kube-api-access-85jw8\") on node \"crc\" DevicePath \"\"" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.563324 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7775a907-216c-477c-9087-4cd2393ee743-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.882407 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" event={"ID":"7775a907-216c-477c-9087-4cd2393ee743","Type":"ContainerDied","Data":"e6c7c4b04e2d8683cdc68e10009afd896b575729885064a14d2bc7f5dcaa0508"} Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.882457 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hhcvg" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.882477 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6c7c4b04e2d8683cdc68e10009afd896b575729885064a14d2bc7f5dcaa0508" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.972318 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-zddbb"] Dec 06 08:48:02 crc kubenswrapper[4945]: E1206 08:48:02.972828 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7775a907-216c-477c-9087-4cd2393ee743" containerName="download-cache-openstack-openstack-cell1" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.972854 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7775a907-216c-477c-9087-4cd2393ee743" containerName="download-cache-openstack-openstack-cell1" Dec 06 08:48:02 crc kubenswrapper[4945]: E1206 08:48:02.972869 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d040649e-1c02-4d63-9f68-861cf49a7013" containerName="registry-server" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.972877 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d040649e-1c02-4d63-9f68-861cf49a7013" containerName="registry-server" Dec 06 08:48:02 crc kubenswrapper[4945]: E1206 08:48:02.972915 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d040649e-1c02-4d63-9f68-861cf49a7013" containerName="extract-content" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.972921 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d040649e-1c02-4d63-9f68-861cf49a7013" containerName="extract-content" Dec 06 08:48:02 crc kubenswrapper[4945]: E1206 08:48:02.972933 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d040649e-1c02-4d63-9f68-861cf49a7013" containerName="extract-utilities" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.972939 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d040649e-1c02-4d63-9f68-861cf49a7013" containerName="extract-utilities" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.973142 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d040649e-1c02-4d63-9f68-861cf49a7013" containerName="registry-server" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.973168 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7775a907-216c-477c-9087-4cd2393ee743" containerName="download-cache-openstack-openstack-cell1" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.974107 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.976029 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.978380 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.978804 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.981067 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:48:02 crc kubenswrapper[4945]: I1206 08:48:02.985719 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-zddbb"] Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.178735 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ssh-key\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.179105 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-inventory\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.179166 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ceph\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.179380 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2c7p\" (UniqueName: \"kubernetes.io/projected/74371052-d51c-4094-be06-38d48cb5053a-kube-api-access-v2c7p\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.280758 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2c7p\" (UniqueName: \"kubernetes.io/projected/74371052-d51c-4094-be06-38d48cb5053a-kube-api-access-v2c7p\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.280889 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ssh-key\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.280932 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-inventory\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.280954 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ceph\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.285720 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-inventory\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.286677 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ssh-key\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.291537 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ceph\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.296976 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2c7p\" (UniqueName: \"kubernetes.io/projected/74371052-d51c-4094-be06-38d48cb5053a-kube-api-access-v2c7p\") pod \"configure-network-openstack-openstack-cell1-zddbb\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.298746 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.837571 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-zddbb"] Dec 06 08:48:03 crc kubenswrapper[4945]: I1206 08:48:03.893689 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-zddbb" event={"ID":"74371052-d51c-4094-be06-38d48cb5053a","Type":"ContainerStarted","Data":"13e5cdde39c4fe762b7dd5311b56fcf077080de8fb7757c70bf7d316997041e7"} Dec 06 08:48:04 crc kubenswrapper[4945]: I1206 08:48:04.904308 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-zddbb" event={"ID":"74371052-d51c-4094-be06-38d48cb5053a","Type":"ContainerStarted","Data":"4cf46103f11d73e7e5f74ad58966569d04ea2244ae055b2c9591b1cf546b131d"} Dec 06 08:48:04 crc kubenswrapper[4945]: I1206 08:48:04.932007 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-zddbb" podStartSLOduration=2.494299711 podStartE2EDuration="2.931983301s" podCreationTimestamp="2025-12-06 08:48:02 +0000 UTC" firstStartedPulling="2025-12-06 08:48:03.843705593 +0000 UTC m=+6897.298566657" lastFinishedPulling="2025-12-06 08:48:04.281389203 +0000 UTC m=+6897.736250247" observedRunningTime="2025-12-06 08:48:04.917840072 +0000 UTC m=+6898.372701126" watchObservedRunningTime="2025-12-06 08:48:04.931983301 +0000 UTC m=+6898.386844345" Dec 06 08:48:08 crc kubenswrapper[4945]: I1206 08:48:08.955780 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:48:08 crc kubenswrapper[4945]: E1206 08:48:08.956533 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:48:20 crc kubenswrapper[4945]: I1206 08:48:20.953263 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:48:20 crc kubenswrapper[4945]: E1206 08:48:20.955476 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:48:31 crc kubenswrapper[4945]: I1206 08:48:31.952949 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:48:31 crc kubenswrapper[4945]: E1206 08:48:31.954600 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:48:43 crc kubenswrapper[4945]: I1206 08:48:43.953652 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:48:43 crc kubenswrapper[4945]: E1206 08:48:43.954479 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:48:54 crc kubenswrapper[4945]: I1206 08:48:54.953983 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:48:54 crc kubenswrapper[4945]: E1206 08:48:54.954931 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.083633 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v7wfc"] Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.086588 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.093149 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v7wfc"] Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.234312 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-utilities\") pod \"certified-operators-v7wfc\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.234484 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn58d\" (UniqueName: \"kubernetes.io/projected/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-kube-api-access-jn58d\") pod \"certified-operators-v7wfc\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.234528 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-catalog-content\") pod \"certified-operators-v7wfc\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.336444 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-utilities\") pod \"certified-operators-v7wfc\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.336614 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn58d\" (UniqueName: \"kubernetes.io/projected/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-kube-api-access-jn58d\") pod \"certified-operators-v7wfc\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.336652 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-catalog-content\") pod \"certified-operators-v7wfc\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.336957 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-utilities\") pod \"certified-operators-v7wfc\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.337000 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-catalog-content\") pod \"certified-operators-v7wfc\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.357806 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn58d\" (UniqueName: \"kubernetes.io/projected/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-kube-api-access-jn58d\") pod \"certified-operators-v7wfc\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.415653 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:05 crc kubenswrapper[4945]: I1206 08:49:05.956798 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v7wfc"] Dec 06 08:49:06 crc kubenswrapper[4945]: I1206 08:49:06.471634 4945 generic.go:334] "Generic (PLEG): container finished" podID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerID="027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6" exitCode=0 Dec 06 08:49:06 crc kubenswrapper[4945]: I1206 08:49:06.471696 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7wfc" event={"ID":"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9","Type":"ContainerDied","Data":"027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6"} Dec 06 08:49:06 crc kubenswrapper[4945]: I1206 08:49:06.472005 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7wfc" event={"ID":"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9","Type":"ContainerStarted","Data":"14e265f3e8bb2105a37752db5d71e1c86915837f6f233e45ed012aadbc218cf6"} Dec 06 08:49:06 crc kubenswrapper[4945]: I1206 08:49:06.959999 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:49:06 crc kubenswrapper[4945]: E1206 08:49:06.960585 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:49:09 crc kubenswrapper[4945]: I1206 08:49:09.499914 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7wfc" event={"ID":"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9","Type":"ContainerStarted","Data":"d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b"} Dec 06 08:49:10 crc kubenswrapper[4945]: I1206 08:49:10.512784 4945 generic.go:334] "Generic (PLEG): container finished" podID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerID="d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b" exitCode=0 Dec 06 08:49:10 crc kubenswrapper[4945]: I1206 08:49:10.512965 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7wfc" event={"ID":"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9","Type":"ContainerDied","Data":"d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b"} Dec 06 08:49:11 crc kubenswrapper[4945]: I1206 08:49:11.524593 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7wfc" event={"ID":"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9","Type":"ContainerStarted","Data":"c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4"} Dec 06 08:49:11 crc kubenswrapper[4945]: I1206 08:49:11.547700 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v7wfc" podStartSLOduration=2.099221816 podStartE2EDuration="6.547678361s" podCreationTimestamp="2025-12-06 08:49:05 +0000 UTC" firstStartedPulling="2025-12-06 08:49:06.475449076 +0000 UTC m=+6959.930310120" lastFinishedPulling="2025-12-06 08:49:10.923905621 +0000 UTC m=+6964.378766665" observedRunningTime="2025-12-06 08:49:11.540610232 +0000 UTC m=+6964.995471326" watchObservedRunningTime="2025-12-06 08:49:11.547678361 +0000 UTC m=+6965.002539415" Dec 06 08:49:15 crc kubenswrapper[4945]: I1206 08:49:15.416383 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:15 crc kubenswrapper[4945]: I1206 08:49:15.416741 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:15 crc kubenswrapper[4945]: I1206 08:49:15.464223 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:18 crc kubenswrapper[4945]: I1206 08:49:18.953238 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:49:18 crc kubenswrapper[4945]: E1206 08:49:18.953850 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:49:21 crc kubenswrapper[4945]: I1206 08:49:21.615432 4945 generic.go:334] "Generic (PLEG): container finished" podID="74371052-d51c-4094-be06-38d48cb5053a" containerID="4cf46103f11d73e7e5f74ad58966569d04ea2244ae055b2c9591b1cf546b131d" exitCode=0 Dec 06 08:49:21 crc kubenswrapper[4945]: I1206 08:49:21.615526 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-zddbb" event={"ID":"74371052-d51c-4094-be06-38d48cb5053a","Type":"ContainerDied","Data":"4cf46103f11d73e7e5f74ad58966569d04ea2244ae055b2c9591b1cf546b131d"} Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.094293 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.217766 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ssh-key\") pod \"74371052-d51c-4094-be06-38d48cb5053a\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.218197 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ceph\") pod \"74371052-d51c-4094-be06-38d48cb5053a\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.218242 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2c7p\" (UniqueName: \"kubernetes.io/projected/74371052-d51c-4094-be06-38d48cb5053a-kube-api-access-v2c7p\") pod \"74371052-d51c-4094-be06-38d48cb5053a\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.218353 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-inventory\") pod \"74371052-d51c-4094-be06-38d48cb5053a\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.224747 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ceph" (OuterVolumeSpecName: "ceph") pod "74371052-d51c-4094-be06-38d48cb5053a" (UID: "74371052-d51c-4094-be06-38d48cb5053a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.226327 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74371052-d51c-4094-be06-38d48cb5053a-kube-api-access-v2c7p" (OuterVolumeSpecName: "kube-api-access-v2c7p") pod "74371052-d51c-4094-be06-38d48cb5053a" (UID: "74371052-d51c-4094-be06-38d48cb5053a"). InnerVolumeSpecName "kube-api-access-v2c7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:49:23 crc kubenswrapper[4945]: E1206 08:49:23.246774 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-inventory podName:74371052-d51c-4094-be06-38d48cb5053a nodeName:}" failed. No retries permitted until 2025-12-06 08:49:23.746732416 +0000 UTC m=+6977.201593460 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-inventory") pod "74371052-d51c-4094-be06-38d48cb5053a" (UID: "74371052-d51c-4094-be06-38d48cb5053a") : error deleting /var/lib/kubelet/pods/74371052-d51c-4094-be06-38d48cb5053a/volume-subpaths: remove /var/lib/kubelet/pods/74371052-d51c-4094-be06-38d48cb5053a/volume-subpaths: no such file or directory Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.249884 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "74371052-d51c-4094-be06-38d48cb5053a" (UID: "74371052-d51c-4094-be06-38d48cb5053a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.320375 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.320441 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.320454 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2c7p\" (UniqueName: \"kubernetes.io/projected/74371052-d51c-4094-be06-38d48cb5053a-kube-api-access-v2c7p\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.644362 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-zddbb" event={"ID":"74371052-d51c-4094-be06-38d48cb5053a","Type":"ContainerDied","Data":"13e5cdde39c4fe762b7dd5311b56fcf077080de8fb7757c70bf7d316997041e7"} Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.644400 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13e5cdde39c4fe762b7dd5311b56fcf077080de8fb7757c70bf7d316997041e7" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.644454 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-zddbb" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.718644 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-gch6j"] Dec 06 08:49:23 crc kubenswrapper[4945]: E1206 08:49:23.719254 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74371052-d51c-4094-be06-38d48cb5053a" containerName="configure-network-openstack-openstack-cell1" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.719279 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="74371052-d51c-4094-be06-38d48cb5053a" containerName="configure-network-openstack-openstack-cell1" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.719586 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="74371052-d51c-4094-be06-38d48cb5053a" containerName="configure-network-openstack-openstack-cell1" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.720866 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.727368 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ceph\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.727457 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ssh-key\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.727648 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-inventory\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.727780 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnfjt\" (UniqueName: \"kubernetes.io/projected/653aa8ac-b5b0-47a6-9901-f6d09ca71674-kube-api-access-fnfjt\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.731352 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-gch6j"] Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.829206 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-inventory\") pod \"74371052-d51c-4094-be06-38d48cb5053a\" (UID: \"74371052-d51c-4094-be06-38d48cb5053a\") " Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.829506 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnfjt\" (UniqueName: \"kubernetes.io/projected/653aa8ac-b5b0-47a6-9901-f6d09ca71674-kube-api-access-fnfjt\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.829625 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ceph\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.829686 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ssh-key\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.829777 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-inventory\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.833399 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-inventory" (OuterVolumeSpecName: "inventory") pod "74371052-d51c-4094-be06-38d48cb5053a" (UID: "74371052-d51c-4094-be06-38d48cb5053a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.833971 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-inventory\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.834002 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ssh-key\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.845468 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ceph\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.847760 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnfjt\" (UniqueName: \"kubernetes.io/projected/653aa8ac-b5b0-47a6-9901-f6d09ca71674-kube-api-access-fnfjt\") pod \"validate-network-openstack-openstack-cell1-gch6j\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:23 crc kubenswrapper[4945]: I1206 08:49:23.932130 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74371052-d51c-4094-be06-38d48cb5053a-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:24 crc kubenswrapper[4945]: I1206 08:49:24.056542 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:24 crc kubenswrapper[4945]: I1206 08:49:24.545230 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-gch6j"] Dec 06 08:49:24 crc kubenswrapper[4945]: I1206 08:49:24.653996 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-gch6j" event={"ID":"653aa8ac-b5b0-47a6-9901-f6d09ca71674","Type":"ContainerStarted","Data":"c0055733b297627fc4caafeb097ee85dea8ecbe1adfc1834c9eaaceff9ef8716"} Dec 06 08:49:25 crc kubenswrapper[4945]: I1206 08:49:25.475626 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:25 crc kubenswrapper[4945]: I1206 08:49:25.523899 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v7wfc"] Dec 06 08:49:25 crc kubenswrapper[4945]: I1206 08:49:25.664038 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-gch6j" event={"ID":"653aa8ac-b5b0-47a6-9901-f6d09ca71674","Type":"ContainerStarted","Data":"21f673b7df219bae36b2655508cd7af065027d1f0d91e9305b5da14d549e8d78"} Dec 06 08:49:25 crc kubenswrapper[4945]: I1206 08:49:25.664217 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v7wfc" podUID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerName="registry-server" containerID="cri-o://c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4" gracePeriod=2 Dec 06 08:49:25 crc kubenswrapper[4945]: I1206 08:49:25.687945 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-gch6j" podStartSLOduration=2.132951226 podStartE2EDuration="2.687920405s" podCreationTimestamp="2025-12-06 08:49:23 +0000 UTC" firstStartedPulling="2025-12-06 08:49:24.547544842 +0000 UTC m=+6978.002405886" lastFinishedPulling="2025-12-06 08:49:25.102514021 +0000 UTC m=+6978.557375065" observedRunningTime="2025-12-06 08:49:25.678564394 +0000 UTC m=+6979.133425438" watchObservedRunningTime="2025-12-06 08:49:25.687920405 +0000 UTC m=+6979.142781449" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.239517 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.378692 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn58d\" (UniqueName: \"kubernetes.io/projected/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-kube-api-access-jn58d\") pod \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.378802 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-utilities\") pod \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.379039 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-catalog-content\") pod \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\" (UID: \"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9\") " Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.379823 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-utilities" (OuterVolumeSpecName: "utilities") pod "3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" (UID: "3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.384413 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-kube-api-access-jn58d" (OuterVolumeSpecName: "kube-api-access-jn58d") pod "3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" (UID: "3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9"). InnerVolumeSpecName "kube-api-access-jn58d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.425572 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" (UID: "3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.481775 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn58d\" (UniqueName: \"kubernetes.io/projected/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-kube-api-access-jn58d\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.481817 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.481830 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.673658 4945 generic.go:334] "Generic (PLEG): container finished" podID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerID="c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4" exitCode=0 Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.673703 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7wfc" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.673730 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7wfc" event={"ID":"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9","Type":"ContainerDied","Data":"c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4"} Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.674118 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7wfc" event={"ID":"3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9","Type":"ContainerDied","Data":"14e265f3e8bb2105a37752db5d71e1c86915837f6f233e45ed012aadbc218cf6"} Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.674142 4945 scope.go:117] "RemoveContainer" containerID="c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.710606 4945 scope.go:117] "RemoveContainer" containerID="d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.714491 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v7wfc"] Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.731758 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v7wfc"] Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.733259 4945 scope.go:117] "RemoveContainer" containerID="027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.790386 4945 scope.go:117] "RemoveContainer" containerID="c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4" Dec 06 08:49:26 crc kubenswrapper[4945]: E1206 08:49:26.790808 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4\": container with ID starting with c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4 not found: ID does not exist" containerID="c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.790851 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4"} err="failed to get container status \"c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4\": rpc error: code = NotFound desc = could not find container \"c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4\": container with ID starting with c5c9e525f2a4bddf5265489092436d13d05e7fa9475c0ecbc736740fa4cba8f4 not found: ID does not exist" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.790876 4945 scope.go:117] "RemoveContainer" containerID="d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b" Dec 06 08:49:26 crc kubenswrapper[4945]: E1206 08:49:26.791178 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b\": container with ID starting with d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b not found: ID does not exist" containerID="d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.791209 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b"} err="failed to get container status \"d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b\": rpc error: code = NotFound desc = could not find container \"d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b\": container with ID starting with d9b0706cfdd109f0b98b017c3b394953890cf9e54d10f79bf45f7ab4e5b85b1b not found: ID does not exist" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.791231 4945 scope.go:117] "RemoveContainer" containerID="027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6" Dec 06 08:49:26 crc kubenswrapper[4945]: E1206 08:49:26.791481 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6\": container with ID starting with 027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6 not found: ID does not exist" containerID="027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.791502 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6"} err="failed to get container status \"027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6\": rpc error: code = NotFound desc = could not find container \"027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6\": container with ID starting with 027b05cf50718a600b84372e3683595caeebf5e3e5b24cdeac4ab860bbad33a6 not found: ID does not exist" Dec 06 08:49:26 crc kubenswrapper[4945]: I1206 08:49:26.965523 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" path="/var/lib/kubelet/pods/3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9/volumes" Dec 06 08:49:30 crc kubenswrapper[4945]: I1206 08:49:30.729380 4945 generic.go:334] "Generic (PLEG): container finished" podID="653aa8ac-b5b0-47a6-9901-f6d09ca71674" containerID="21f673b7df219bae36b2655508cd7af065027d1f0d91e9305b5da14d549e8d78" exitCode=0 Dec 06 08:49:30 crc kubenswrapper[4945]: I1206 08:49:30.729494 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-gch6j" event={"ID":"653aa8ac-b5b0-47a6-9901-f6d09ca71674","Type":"ContainerDied","Data":"21f673b7df219bae36b2655508cd7af065027d1f0d91e9305b5da14d549e8d78"} Dec 06 08:49:30 crc kubenswrapper[4945]: I1206 08:49:30.953580 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:49:30 crc kubenswrapper[4945]: E1206 08:49:30.953811 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.278249 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.399206 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ceph\") pod \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.399572 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ssh-key\") pod \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.399777 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnfjt\" (UniqueName: \"kubernetes.io/projected/653aa8ac-b5b0-47a6-9901-f6d09ca71674-kube-api-access-fnfjt\") pod \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.399841 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-inventory\") pod \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\" (UID: \"653aa8ac-b5b0-47a6-9901-f6d09ca71674\") " Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.405830 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/653aa8ac-b5b0-47a6-9901-f6d09ca71674-kube-api-access-fnfjt" (OuterVolumeSpecName: "kube-api-access-fnfjt") pod "653aa8ac-b5b0-47a6-9901-f6d09ca71674" (UID: "653aa8ac-b5b0-47a6-9901-f6d09ca71674"). InnerVolumeSpecName "kube-api-access-fnfjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.406181 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ceph" (OuterVolumeSpecName: "ceph") pod "653aa8ac-b5b0-47a6-9901-f6d09ca71674" (UID: "653aa8ac-b5b0-47a6-9901-f6d09ca71674"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.428434 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-inventory" (OuterVolumeSpecName: "inventory") pod "653aa8ac-b5b0-47a6-9901-f6d09ca71674" (UID: "653aa8ac-b5b0-47a6-9901-f6d09ca71674"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.429281 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "653aa8ac-b5b0-47a6-9901-f6d09ca71674" (UID: "653aa8ac-b5b0-47a6-9901-f6d09ca71674"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.502092 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnfjt\" (UniqueName: \"kubernetes.io/projected/653aa8ac-b5b0-47a6-9901-f6d09ca71674-kube-api-access-fnfjt\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.502136 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.502147 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.502157 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/653aa8ac-b5b0-47a6-9901-f6d09ca71674-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.748763 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-gch6j" event={"ID":"653aa8ac-b5b0-47a6-9901-f6d09ca71674","Type":"ContainerDied","Data":"c0055733b297627fc4caafeb097ee85dea8ecbe1adfc1834c9eaaceff9ef8716"} Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.748812 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0055733b297627fc4caafeb097ee85dea8ecbe1adfc1834c9eaaceff9ef8716" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.748887 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-gch6j" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.822503 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-vv8q2"] Dec 06 08:49:32 crc kubenswrapper[4945]: E1206 08:49:32.822929 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerName="extract-content" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.822947 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerName="extract-content" Dec 06 08:49:32 crc kubenswrapper[4945]: E1206 08:49:32.822985 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerName="extract-utilities" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.822992 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerName="extract-utilities" Dec 06 08:49:32 crc kubenswrapper[4945]: E1206 08:49:32.823008 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="653aa8ac-b5b0-47a6-9901-f6d09ca71674" containerName="validate-network-openstack-openstack-cell1" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.823015 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="653aa8ac-b5b0-47a6-9901-f6d09ca71674" containerName="validate-network-openstack-openstack-cell1" Dec 06 08:49:32 crc kubenswrapper[4945]: E1206 08:49:32.823032 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerName="registry-server" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.823038 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerName="registry-server" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.823228 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eb6dbd1-bac5-4a69-8c18-db3806c2a6c9" containerName="registry-server" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.823249 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="653aa8ac-b5b0-47a6-9901-f6d09ca71674" containerName="validate-network-openstack-openstack-cell1" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.823955 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.830518 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.830537 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.830946 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.831005 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.831592 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-vv8q2"] Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.909560 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-inventory\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.909632 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc7mz\" (UniqueName: \"kubernetes.io/projected/c28069dc-d3fb-40e7-8235-72c59c6580d8-kube-api-access-rc7mz\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.909826 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ssh-key\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:32 crc kubenswrapper[4945]: I1206 08:49:32.910017 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ceph\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.011502 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ssh-key\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.011582 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ceph\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.011766 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-inventory\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.011812 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc7mz\" (UniqueName: \"kubernetes.io/projected/c28069dc-d3fb-40e7-8235-72c59c6580d8-kube-api-access-rc7mz\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.024560 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-inventory\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.024569 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ssh-key\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.024678 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ceph\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.028026 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc7mz\" (UniqueName: \"kubernetes.io/projected/c28069dc-d3fb-40e7-8235-72c59c6580d8-kube-api-access-rc7mz\") pod \"install-os-openstack-openstack-cell1-vv8q2\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.141665 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.640204 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-vv8q2"] Dec 06 08:49:33 crc kubenswrapper[4945]: I1206 08:49:33.762489 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vv8q2" event={"ID":"c28069dc-d3fb-40e7-8235-72c59c6580d8","Type":"ContainerStarted","Data":"7933287c6d9f93641f56dfaa2db37adce547d395d64e3097e5a10bdb5cbbc74a"} Dec 06 08:49:34 crc kubenswrapper[4945]: I1206 08:49:34.771635 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vv8q2" event={"ID":"c28069dc-d3fb-40e7-8235-72c59c6580d8","Type":"ContainerStarted","Data":"63b7bd9c1214f4dcbe37e69b71b192887867ed9dd87af7ba83c3b178172cca76"} Dec 06 08:49:34 crc kubenswrapper[4945]: I1206 08:49:34.791887 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-vv8q2" podStartSLOduration=2.372007471 podStartE2EDuration="2.791870284s" podCreationTimestamp="2025-12-06 08:49:32 +0000 UTC" firstStartedPulling="2025-12-06 08:49:33.647256729 +0000 UTC m=+6987.102117793" lastFinishedPulling="2025-12-06 08:49:34.067119562 +0000 UTC m=+6987.521980606" observedRunningTime="2025-12-06 08:49:34.78722047 +0000 UTC m=+6988.242081514" watchObservedRunningTime="2025-12-06 08:49:34.791870284 +0000 UTC m=+6988.246731328" Dec 06 08:49:42 crc kubenswrapper[4945]: I1206 08:49:42.954842 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:49:42 crc kubenswrapper[4945]: E1206 08:49:42.955818 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:49:57 crc kubenswrapper[4945]: I1206 08:49:57.953936 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:49:57 crc kubenswrapper[4945]: E1206 08:49:57.954836 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:50:12 crc kubenswrapper[4945]: I1206 08:50:12.953527 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:50:14 crc kubenswrapper[4945]: I1206 08:50:14.119858 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"77f6ccf1b6f2295e7b8cc956a5858de679eeb654affbd183de5ef1d7218d8eeb"} Dec 06 08:50:18 crc kubenswrapper[4945]: I1206 08:50:18.176546 4945 generic.go:334] "Generic (PLEG): container finished" podID="c28069dc-d3fb-40e7-8235-72c59c6580d8" containerID="63b7bd9c1214f4dcbe37e69b71b192887867ed9dd87af7ba83c3b178172cca76" exitCode=0 Dec 06 08:50:18 crc kubenswrapper[4945]: I1206 08:50:18.177085 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vv8q2" event={"ID":"c28069dc-d3fb-40e7-8235-72c59c6580d8","Type":"ContainerDied","Data":"63b7bd9c1214f4dcbe37e69b71b192887867ed9dd87af7ba83c3b178172cca76"} Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.640526 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.723076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ssh-key\") pod \"c28069dc-d3fb-40e7-8235-72c59c6580d8\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.723126 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-inventory\") pod \"c28069dc-d3fb-40e7-8235-72c59c6580d8\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.723266 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc7mz\" (UniqueName: \"kubernetes.io/projected/c28069dc-d3fb-40e7-8235-72c59c6580d8-kube-api-access-rc7mz\") pod \"c28069dc-d3fb-40e7-8235-72c59c6580d8\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.723420 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ceph\") pod \"c28069dc-d3fb-40e7-8235-72c59c6580d8\" (UID: \"c28069dc-d3fb-40e7-8235-72c59c6580d8\") " Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.730984 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ceph" (OuterVolumeSpecName: "ceph") pod "c28069dc-d3fb-40e7-8235-72c59c6580d8" (UID: "c28069dc-d3fb-40e7-8235-72c59c6580d8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.731232 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c28069dc-d3fb-40e7-8235-72c59c6580d8-kube-api-access-rc7mz" (OuterVolumeSpecName: "kube-api-access-rc7mz") pod "c28069dc-d3fb-40e7-8235-72c59c6580d8" (UID: "c28069dc-d3fb-40e7-8235-72c59c6580d8"). InnerVolumeSpecName "kube-api-access-rc7mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.762653 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-inventory" (OuterVolumeSpecName: "inventory") pod "c28069dc-d3fb-40e7-8235-72c59c6580d8" (UID: "c28069dc-d3fb-40e7-8235-72c59c6580d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.781941 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c28069dc-d3fb-40e7-8235-72c59c6580d8" (UID: "c28069dc-d3fb-40e7-8235-72c59c6580d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.826113 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.826451 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.826520 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c28069dc-d3fb-40e7-8235-72c59c6580d8-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:50:19 crc kubenswrapper[4945]: I1206 08:50:19.826585 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc7mz\" (UniqueName: \"kubernetes.io/projected/c28069dc-d3fb-40e7-8235-72c59c6580d8-kube-api-access-rc7mz\") on node \"crc\" DevicePath \"\"" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.197968 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vv8q2" event={"ID":"c28069dc-d3fb-40e7-8235-72c59c6580d8","Type":"ContainerDied","Data":"7933287c6d9f93641f56dfaa2db37adce547d395d64e3097e5a10bdb5cbbc74a"} Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.198242 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7933287c6d9f93641f56dfaa2db37adce547d395d64e3097e5a10bdb5cbbc74a" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.198070 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vv8q2" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.284982 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-fxlf6"] Dec 06 08:50:20 crc kubenswrapper[4945]: E1206 08:50:20.285393 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c28069dc-d3fb-40e7-8235-72c59c6580d8" containerName="install-os-openstack-openstack-cell1" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.285410 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c28069dc-d3fb-40e7-8235-72c59c6580d8" containerName="install-os-openstack-openstack-cell1" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.285624 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c28069dc-d3fb-40e7-8235-72c59c6580d8" containerName="install-os-openstack-openstack-cell1" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.286305 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.289741 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.289828 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.289845 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.291106 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.306334 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-fxlf6"] Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.440067 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-inventory\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.440112 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ssh-key\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.440254 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb896\" (UniqueName: \"kubernetes.io/projected/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-kube-api-access-tb896\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.440490 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ceph\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.542219 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ceph\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.542343 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-inventory\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.542364 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ssh-key\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.542401 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb896\" (UniqueName: \"kubernetes.io/projected/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-kube-api-access-tb896\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.546694 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ssh-key\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.546714 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-inventory\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.553962 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ceph\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.561587 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb896\" (UniqueName: \"kubernetes.io/projected/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-kube-api-access-tb896\") pod \"configure-os-openstack-openstack-cell1-fxlf6\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:20 crc kubenswrapper[4945]: I1206 08:50:20.611089 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:50:21 crc kubenswrapper[4945]: I1206 08:50:21.141608 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-fxlf6"] Dec 06 08:50:21 crc kubenswrapper[4945]: I1206 08:50:21.209620 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" event={"ID":"92e071d6-dcaf-49d5-8dc0-3d8b3264b585","Type":"ContainerStarted","Data":"3fff7c5060e48910dc469c47eba7bcff958cdfa6a6eadc5768f58c4c0f554c78"} Dec 06 08:50:22 crc kubenswrapper[4945]: I1206 08:50:22.217731 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" event={"ID":"92e071d6-dcaf-49d5-8dc0-3d8b3264b585","Type":"ContainerStarted","Data":"9f96eb601a80839193909cf5b3a8a1af5c7abf3d969e142b082de65bcd73a08c"} Dec 06 08:50:22 crc kubenswrapper[4945]: I1206 08:50:22.239881 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" podStartSLOduration=1.8032732870000001 podStartE2EDuration="2.239859958s" podCreationTimestamp="2025-12-06 08:50:20 +0000 UTC" firstStartedPulling="2025-12-06 08:50:21.140910424 +0000 UTC m=+7034.595771468" lastFinishedPulling="2025-12-06 08:50:21.577497095 +0000 UTC m=+7035.032358139" observedRunningTime="2025-12-06 08:50:22.235133762 +0000 UTC m=+7035.689994816" watchObservedRunningTime="2025-12-06 08:50:22.239859958 +0000 UTC m=+7035.694721002" Dec 06 08:51:04 crc kubenswrapper[4945]: I1206 08:51:04.822712 4945 generic.go:334] "Generic (PLEG): container finished" podID="92e071d6-dcaf-49d5-8dc0-3d8b3264b585" containerID="9f96eb601a80839193909cf5b3a8a1af5c7abf3d969e142b082de65bcd73a08c" exitCode=0 Dec 06 08:51:04 crc kubenswrapper[4945]: I1206 08:51:04.822825 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" event={"ID":"92e071d6-dcaf-49d5-8dc0-3d8b3264b585","Type":"ContainerDied","Data":"9f96eb601a80839193909cf5b3a8a1af5c7abf3d969e142b082de65bcd73a08c"} Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.359430 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.436389 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ssh-key\") pod \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.436483 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-inventory\") pod \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.436702 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb896\" (UniqueName: \"kubernetes.io/projected/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-kube-api-access-tb896\") pod \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.436843 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ceph\") pod \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\" (UID: \"92e071d6-dcaf-49d5-8dc0-3d8b3264b585\") " Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.442694 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-kube-api-access-tb896" (OuterVolumeSpecName: "kube-api-access-tb896") pod "92e071d6-dcaf-49d5-8dc0-3d8b3264b585" (UID: "92e071d6-dcaf-49d5-8dc0-3d8b3264b585"). InnerVolumeSpecName "kube-api-access-tb896". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.442888 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ceph" (OuterVolumeSpecName: "ceph") pod "92e071d6-dcaf-49d5-8dc0-3d8b3264b585" (UID: "92e071d6-dcaf-49d5-8dc0-3d8b3264b585"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.465417 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-inventory" (OuterVolumeSpecName: "inventory") pod "92e071d6-dcaf-49d5-8dc0-3d8b3264b585" (UID: "92e071d6-dcaf-49d5-8dc0-3d8b3264b585"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.469841 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "92e071d6-dcaf-49d5-8dc0-3d8b3264b585" (UID: "92e071d6-dcaf-49d5-8dc0-3d8b3264b585"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.539703 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.539762 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.539775 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.539789 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb896\" (UniqueName: \"kubernetes.io/projected/92e071d6-dcaf-49d5-8dc0-3d8b3264b585-kube-api-access-tb896\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.841054 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" event={"ID":"92e071d6-dcaf-49d5-8dc0-3d8b3264b585","Type":"ContainerDied","Data":"3fff7c5060e48910dc469c47eba7bcff958cdfa6a6eadc5768f58c4c0f554c78"} Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.841097 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fff7c5060e48910dc469c47eba7bcff958cdfa6a6eadc5768f58c4c0f554c78" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.841118 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-fxlf6" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.933152 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-skqg9"] Dec 06 08:51:06 crc kubenswrapper[4945]: E1206 08:51:06.933744 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92e071d6-dcaf-49d5-8dc0-3d8b3264b585" containerName="configure-os-openstack-openstack-cell1" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.933771 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="92e071d6-dcaf-49d5-8dc0-3d8b3264b585" containerName="configure-os-openstack-openstack-cell1" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.934033 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="92e071d6-dcaf-49d5-8dc0-3d8b3264b585" containerName="configure-os-openstack-openstack-cell1" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.935575 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.937718 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.937952 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.938082 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.938245 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:51:06 crc kubenswrapper[4945]: I1206 08:51:06.945166 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-skqg9"] Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.050352 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ceph\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.050432 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-inventory-0\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.050469 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.050631 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzkbm\" (UniqueName: \"kubernetes.io/projected/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-kube-api-access-tzkbm\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.152204 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.152723 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzkbm\" (UniqueName: \"kubernetes.io/projected/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-kube-api-access-tzkbm\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.152769 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ceph\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.152917 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-inventory-0\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.155485 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.155778 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.156453 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ceph\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.165897 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.169174 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzkbm\" (UniqueName: \"kubernetes.io/projected/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-kube-api-access-tzkbm\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.179766 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-inventory-0\") pod \"ssh-known-hosts-openstack-skqg9\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.255108 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.263423 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.791125 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-skqg9"] Dec 06 08:51:07 crc kubenswrapper[4945]: I1206 08:51:07.849716 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-skqg9" event={"ID":"aa10039e-9b12-42f0-bdf9-b91b3e7acf10","Type":"ContainerStarted","Data":"841da33b9a7f495c2645fbccf2e6424e97ce59084f7b6cdb4b19eab30b5c4010"} Dec 06 08:51:08 crc kubenswrapper[4945]: I1206 08:51:08.207622 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:51:08 crc kubenswrapper[4945]: I1206 08:51:08.859292 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-skqg9" event={"ID":"aa10039e-9b12-42f0-bdf9-b91b3e7acf10","Type":"ContainerStarted","Data":"5fa8e72f5a479b72b5e2570845ca05905f957899e0128604574a65d7fe1052e8"} Dec 06 08:51:08 crc kubenswrapper[4945]: I1206 08:51:08.884579 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-skqg9" podStartSLOduration=2.477259702 podStartE2EDuration="2.88456202s" podCreationTimestamp="2025-12-06 08:51:06 +0000 UTC" firstStartedPulling="2025-12-06 08:51:07.796155948 +0000 UTC m=+7081.251016992" lastFinishedPulling="2025-12-06 08:51:08.203458266 +0000 UTC m=+7081.658319310" observedRunningTime="2025-12-06 08:51:08.876622138 +0000 UTC m=+7082.331483192" watchObservedRunningTime="2025-12-06 08:51:08.88456202 +0000 UTC m=+7082.339423064" Dec 06 08:51:16 crc kubenswrapper[4945]: I1206 08:51:16.975487 4945 generic.go:334] "Generic (PLEG): container finished" podID="aa10039e-9b12-42f0-bdf9-b91b3e7acf10" containerID="5fa8e72f5a479b72b5e2570845ca05905f957899e0128604574a65d7fe1052e8" exitCode=0 Dec 06 08:51:16 crc kubenswrapper[4945]: I1206 08:51:16.977535 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-skqg9" event={"ID":"aa10039e-9b12-42f0-bdf9-b91b3e7acf10","Type":"ContainerDied","Data":"5fa8e72f5a479b72b5e2570845ca05905f957899e0128604574a65d7fe1052e8"} Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.442265 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.490013 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzkbm\" (UniqueName: \"kubernetes.io/projected/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-kube-api-access-tzkbm\") pod \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.490082 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ceph\") pod \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.490233 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ssh-key-openstack-cell1\") pod \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.490355 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-inventory-0\") pod \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\" (UID: \"aa10039e-9b12-42f0-bdf9-b91b3e7acf10\") " Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.496500 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ceph" (OuterVolumeSpecName: "ceph") pod "aa10039e-9b12-42f0-bdf9-b91b3e7acf10" (UID: "aa10039e-9b12-42f0-bdf9-b91b3e7acf10"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.496527 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-kube-api-access-tzkbm" (OuterVolumeSpecName: "kube-api-access-tzkbm") pod "aa10039e-9b12-42f0-bdf9-b91b3e7acf10" (UID: "aa10039e-9b12-42f0-bdf9-b91b3e7acf10"). InnerVolumeSpecName "kube-api-access-tzkbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.521937 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "aa10039e-9b12-42f0-bdf9-b91b3e7acf10" (UID: "aa10039e-9b12-42f0-bdf9-b91b3e7acf10"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.526121 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "aa10039e-9b12-42f0-bdf9-b91b3e7acf10" (UID: "aa10039e-9b12-42f0-bdf9-b91b3e7acf10"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.592970 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzkbm\" (UniqueName: \"kubernetes.io/projected/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-kube-api-access-tzkbm\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.593002 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.593012 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.593021 4945 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/aa10039e-9b12-42f0-bdf9-b91b3e7acf10-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.995300 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-skqg9" event={"ID":"aa10039e-9b12-42f0-bdf9-b91b3e7acf10","Type":"ContainerDied","Data":"841da33b9a7f495c2645fbccf2e6424e97ce59084f7b6cdb4b19eab30b5c4010"} Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.995352 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="841da33b9a7f495c2645fbccf2e6424e97ce59084f7b6cdb4b19eab30b5c4010" Dec 06 08:51:18 crc kubenswrapper[4945]: I1206 08:51:18.995360 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-skqg9" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.074865 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-qtb45"] Dec 06 08:51:19 crc kubenswrapper[4945]: E1206 08:51:19.075320 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa10039e-9b12-42f0-bdf9-b91b3e7acf10" containerName="ssh-known-hosts-openstack" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.075339 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa10039e-9b12-42f0-bdf9-b91b3e7acf10" containerName="ssh-known-hosts-openstack" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.075552 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa10039e-9b12-42f0-bdf9-b91b3e7acf10" containerName="ssh-known-hosts-openstack" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.076330 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.078751 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.078845 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.080610 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.081973 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.096348 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-qtb45"] Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.104024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x84h\" (UniqueName: \"kubernetes.io/projected/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-kube-api-access-6x84h\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.104119 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ssh-key\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.104192 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ceph\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.104263 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-inventory\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.206209 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ssh-key\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.206386 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ceph\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.206515 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-inventory\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.206565 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x84h\" (UniqueName: \"kubernetes.io/projected/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-kube-api-access-6x84h\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.209907 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-inventory\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.209929 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ceph\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.215813 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ssh-key\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.248963 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x84h\" (UniqueName: \"kubernetes.io/projected/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-kube-api-access-6x84h\") pod \"run-os-openstack-openstack-cell1-qtb45\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.395265 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:19 crc kubenswrapper[4945]: I1206 08:51:19.924655 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-qtb45"] Dec 06 08:51:20 crc kubenswrapper[4945]: I1206 08:51:20.006823 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-qtb45" event={"ID":"fc8f860e-c526-4d6a-b91b-6a6ed10e176b","Type":"ContainerStarted","Data":"748316ab5667120133de8c79992b91f6c7f45941e8189a4c3ffa93ebbd60d64e"} Dec 06 08:51:21 crc kubenswrapper[4945]: I1206 08:51:21.018192 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-qtb45" event={"ID":"fc8f860e-c526-4d6a-b91b-6a6ed10e176b","Type":"ContainerStarted","Data":"c27b0448330d100a0433b8e86a95a3a76786b683476d34ba995665aa225f57e6"} Dec 06 08:51:21 crc kubenswrapper[4945]: I1206 08:51:21.052946 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-qtb45" podStartSLOduration=1.6623469530000001 podStartE2EDuration="2.052927374s" podCreationTimestamp="2025-12-06 08:51:19 +0000 UTC" firstStartedPulling="2025-12-06 08:51:19.931274332 +0000 UTC m=+7093.386135376" lastFinishedPulling="2025-12-06 08:51:20.321854753 +0000 UTC m=+7093.776715797" observedRunningTime="2025-12-06 08:51:21.030975916 +0000 UTC m=+7094.485836960" watchObservedRunningTime="2025-12-06 08:51:21.052927374 +0000 UTC m=+7094.507788418" Dec 06 08:51:28 crc kubenswrapper[4945]: I1206 08:51:28.083428 4945 generic.go:334] "Generic (PLEG): container finished" podID="fc8f860e-c526-4d6a-b91b-6a6ed10e176b" containerID="c27b0448330d100a0433b8e86a95a3a76786b683476d34ba995665aa225f57e6" exitCode=0 Dec 06 08:51:28 crc kubenswrapper[4945]: I1206 08:51:28.083715 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-qtb45" event={"ID":"fc8f860e-c526-4d6a-b91b-6a6ed10e176b","Type":"ContainerDied","Data":"c27b0448330d100a0433b8e86a95a3a76786b683476d34ba995665aa225f57e6"} Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.570352 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.625448 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ssh-key\") pod \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.625661 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ceph\") pod \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.625703 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x84h\" (UniqueName: \"kubernetes.io/projected/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-kube-api-access-6x84h\") pod \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.625732 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-inventory\") pod \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\" (UID: \"fc8f860e-c526-4d6a-b91b-6a6ed10e176b\") " Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.632101 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-kube-api-access-6x84h" (OuterVolumeSpecName: "kube-api-access-6x84h") pod "fc8f860e-c526-4d6a-b91b-6a6ed10e176b" (UID: "fc8f860e-c526-4d6a-b91b-6a6ed10e176b"). InnerVolumeSpecName "kube-api-access-6x84h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.639565 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ceph" (OuterVolumeSpecName: "ceph") pod "fc8f860e-c526-4d6a-b91b-6a6ed10e176b" (UID: "fc8f860e-c526-4d6a-b91b-6a6ed10e176b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.663257 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-inventory" (OuterVolumeSpecName: "inventory") pod "fc8f860e-c526-4d6a-b91b-6a6ed10e176b" (UID: "fc8f860e-c526-4d6a-b91b-6a6ed10e176b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.665777 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fc8f860e-c526-4d6a-b91b-6a6ed10e176b" (UID: "fc8f860e-c526-4d6a-b91b-6a6ed10e176b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.728166 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.728407 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x84h\" (UniqueName: \"kubernetes.io/projected/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-kube-api-access-6x84h\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.728425 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:29 crc kubenswrapper[4945]: I1206 08:51:29.728437 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8f860e-c526-4d6a-b91b-6a6ed10e176b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.120873 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-qtb45" event={"ID":"fc8f860e-c526-4d6a-b91b-6a6ed10e176b","Type":"ContainerDied","Data":"748316ab5667120133de8c79992b91f6c7f45941e8189a4c3ffa93ebbd60d64e"} Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.120944 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="748316ab5667120133de8c79992b91f6c7f45941e8189a4c3ffa93ebbd60d64e" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.121036 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-qtb45" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.233891 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-bkpcr"] Dec 06 08:51:30 crc kubenswrapper[4945]: E1206 08:51:30.234360 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc8f860e-c526-4d6a-b91b-6a6ed10e176b" containerName="run-os-openstack-openstack-cell1" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.234376 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc8f860e-c526-4d6a-b91b-6a6ed10e176b" containerName="run-os-openstack-openstack-cell1" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.234564 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc8f860e-c526-4d6a-b91b-6a6ed10e176b" containerName="run-os-openstack-openstack-cell1" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.235256 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.238935 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.238998 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.239134 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.239169 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.261069 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-bkpcr"] Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.341924 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ceph\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.342417 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.342456 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8kf5\" (UniqueName: \"kubernetes.io/projected/74f5fed5-012c-44bd-8e9c-4576689abe4c-kube-api-access-l8kf5\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.342494 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-inventory\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.444583 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-inventory\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.444691 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ceph\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.444811 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.444838 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8kf5\" (UniqueName: \"kubernetes.io/projected/74f5fed5-012c-44bd-8e9c-4576689abe4c-kube-api-access-l8kf5\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.453506 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.453521 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ceph\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.453679 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-inventory\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.462692 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8kf5\" (UniqueName: \"kubernetes.io/projected/74f5fed5-012c-44bd-8e9c-4576689abe4c-kube-api-access-l8kf5\") pod \"reboot-os-openstack-openstack-cell1-bkpcr\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:30 crc kubenswrapper[4945]: I1206 08:51:30.553047 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:31 crc kubenswrapper[4945]: I1206 08:51:31.084454 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-bkpcr"] Dec 06 08:51:31 crc kubenswrapper[4945]: I1206 08:51:31.086945 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 08:51:31 crc kubenswrapper[4945]: I1206 08:51:31.131546 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" event={"ID":"74f5fed5-012c-44bd-8e9c-4576689abe4c","Type":"ContainerStarted","Data":"8fd226bebfcfc07016ae40f0852f816f3978dc9b6554dae1c1a10b7ac78c6b15"} Dec 06 08:51:32 crc kubenswrapper[4945]: I1206 08:51:32.141237 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" event={"ID":"74f5fed5-012c-44bd-8e9c-4576689abe4c","Type":"ContainerStarted","Data":"5360d39eded5d2ab221c52e5515a97b3e7f046b1f72f661011223de83b67b692"} Dec 06 08:51:32 crc kubenswrapper[4945]: I1206 08:51:32.165056 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" podStartSLOduration=1.694031321 podStartE2EDuration="2.165036104s" podCreationTimestamp="2025-12-06 08:51:30 +0000 UTC" firstStartedPulling="2025-12-06 08:51:31.086706222 +0000 UTC m=+7104.541567266" lastFinishedPulling="2025-12-06 08:51:31.557711005 +0000 UTC m=+7105.012572049" observedRunningTime="2025-12-06 08:51:32.158613272 +0000 UTC m=+7105.613474336" watchObservedRunningTime="2025-12-06 08:51:32.165036104 +0000 UTC m=+7105.619897138" Dec 06 08:51:48 crc kubenswrapper[4945]: I1206 08:51:48.343257 4945 generic.go:334] "Generic (PLEG): container finished" podID="74f5fed5-012c-44bd-8e9c-4576689abe4c" containerID="5360d39eded5d2ab221c52e5515a97b3e7f046b1f72f661011223de83b67b692" exitCode=0 Dec 06 08:51:48 crc kubenswrapper[4945]: I1206 08:51:48.343440 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" event={"ID":"74f5fed5-012c-44bd-8e9c-4576689abe4c","Type":"ContainerDied","Data":"5360d39eded5d2ab221c52e5515a97b3e7f046b1f72f661011223de83b67b692"} Dec 06 08:51:49 crc kubenswrapper[4945]: I1206 08:51:49.900964 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:49 crc kubenswrapper[4945]: I1206 08:51:49.989915 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ssh-key\") pod \"74f5fed5-012c-44bd-8e9c-4576689abe4c\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " Dec 06 08:51:49 crc kubenswrapper[4945]: I1206 08:51:49.990215 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-inventory\") pod \"74f5fed5-012c-44bd-8e9c-4576689abe4c\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " Dec 06 08:51:49 crc kubenswrapper[4945]: I1206 08:51:49.990420 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ceph\") pod \"74f5fed5-012c-44bd-8e9c-4576689abe4c\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " Dec 06 08:51:49 crc kubenswrapper[4945]: I1206 08:51:49.990546 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8kf5\" (UniqueName: \"kubernetes.io/projected/74f5fed5-012c-44bd-8e9c-4576689abe4c-kube-api-access-l8kf5\") pod \"74f5fed5-012c-44bd-8e9c-4576689abe4c\" (UID: \"74f5fed5-012c-44bd-8e9c-4576689abe4c\") " Dec 06 08:51:49 crc kubenswrapper[4945]: I1206 08:51:49.996546 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ceph" (OuterVolumeSpecName: "ceph") pod "74f5fed5-012c-44bd-8e9c-4576689abe4c" (UID: "74f5fed5-012c-44bd-8e9c-4576689abe4c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:49.998975 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74f5fed5-012c-44bd-8e9c-4576689abe4c-kube-api-access-l8kf5" (OuterVolumeSpecName: "kube-api-access-l8kf5") pod "74f5fed5-012c-44bd-8e9c-4576689abe4c" (UID: "74f5fed5-012c-44bd-8e9c-4576689abe4c"). InnerVolumeSpecName "kube-api-access-l8kf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.023266 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-inventory" (OuterVolumeSpecName: "inventory") pod "74f5fed5-012c-44bd-8e9c-4576689abe4c" (UID: "74f5fed5-012c-44bd-8e9c-4576689abe4c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.034702 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "74f5fed5-012c-44bd-8e9c-4576689abe4c" (UID: "74f5fed5-012c-44bd-8e9c-4576689abe4c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.093249 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.093311 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8kf5\" (UniqueName: \"kubernetes.io/projected/74f5fed5-012c-44bd-8e9c-4576689abe4c-kube-api-access-l8kf5\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.093328 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.093340 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74f5fed5-012c-44bd-8e9c-4576689abe4c-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.362826 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" event={"ID":"74f5fed5-012c-44bd-8e9c-4576689abe4c","Type":"ContainerDied","Data":"8fd226bebfcfc07016ae40f0852f816f3978dc9b6554dae1c1a10b7ac78c6b15"} Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.362870 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fd226bebfcfc07016ae40f0852f816f3978dc9b6554dae1c1a10b7ac78c6b15" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.362926 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-bkpcr" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.491914 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-gnccs"] Dec 06 08:51:50 crc kubenswrapper[4945]: E1206 08:51:50.492556 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74f5fed5-012c-44bd-8e9c-4576689abe4c" containerName="reboot-os-openstack-openstack-cell1" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.492637 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="74f5fed5-012c-44bd-8e9c-4576689abe4c" containerName="reboot-os-openstack-openstack-cell1" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.492921 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="74f5fed5-012c-44bd-8e9c-4576689abe4c" containerName="reboot-os-openstack-openstack-cell1" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.493803 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.497135 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.497557 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.497712 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.501859 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.506874 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-gnccs"] Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.607586 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.607902 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.608066 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.608166 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.608309 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ceph\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.608427 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-inventory\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.608598 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.608698 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ssh-key\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.608816 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.608914 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.609041 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.609164 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2np7j\" (UniqueName: \"kubernetes.io/projected/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-kube-api-access-2np7j\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.711397 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.711706 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.711851 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.711943 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2np7j\" (UniqueName: \"kubernetes.io/projected/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-kube-api-access-2np7j\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.712034 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.712114 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.712236 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.712362 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.712471 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ceph\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.712587 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-inventory\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.712741 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.712828 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ssh-key\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.722332 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.722993 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.727007 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.727194 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.727828 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.731542 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.732003 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.732336 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ceph\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.732753 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ssh-key\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.733851 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.735666 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-inventory\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.739140 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2np7j\" (UniqueName: \"kubernetes.io/projected/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-kube-api-access-2np7j\") pod \"install-certs-openstack-openstack-cell1-gnccs\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:50 crc kubenswrapper[4945]: I1206 08:51:50.820011 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:51:51 crc kubenswrapper[4945]: W1206 08:51:51.386218 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fcc9e05_4e5d_434c_bd5b_382d452fc94c.slice/crio-28cf1370b46b43f6755383c2de4b76581c3d380bb41072ee07cfbc497c9f2f88 WatchSource:0}: Error finding container 28cf1370b46b43f6755383c2de4b76581c3d380bb41072ee07cfbc497c9f2f88: Status 404 returned error can't find the container with id 28cf1370b46b43f6755383c2de4b76581c3d380bb41072ee07cfbc497c9f2f88 Dec 06 08:51:51 crc kubenswrapper[4945]: I1206 08:51:51.388085 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-gnccs"] Dec 06 08:51:52 crc kubenswrapper[4945]: I1206 08:51:52.384477 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-gnccs" event={"ID":"6fcc9e05-4e5d-434c-bd5b-382d452fc94c","Type":"ContainerStarted","Data":"36897258927ac269a46833c77a9d9c171d86738d060ef81041f0a3c5f4b7c0b6"} Dec 06 08:51:52 crc kubenswrapper[4945]: I1206 08:51:52.384809 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-gnccs" event={"ID":"6fcc9e05-4e5d-434c-bd5b-382d452fc94c","Type":"ContainerStarted","Data":"28cf1370b46b43f6755383c2de4b76581c3d380bb41072ee07cfbc497c9f2f88"} Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.017389 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-gnccs" podStartSLOduration=5.475201603 podStartE2EDuration="6.01737098s" podCreationTimestamp="2025-12-06 08:51:50 +0000 UTC" firstStartedPulling="2025-12-06 08:51:51.388963649 +0000 UTC m=+7124.843824693" lastFinishedPulling="2025-12-06 08:51:51.931133026 +0000 UTC m=+7125.385994070" observedRunningTime="2025-12-06 08:51:52.406935157 +0000 UTC m=+7125.861796211" watchObservedRunningTime="2025-12-06 08:51:56.01737098 +0000 UTC m=+7129.472232024" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.023123 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tm7xt"] Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.025356 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.031942 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-catalog-content\") pod \"redhat-marketplace-tm7xt\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.032149 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k49rw\" (UniqueName: \"kubernetes.io/projected/b5752df4-d789-4fee-a67e-5329584c5120-kube-api-access-k49rw\") pod \"redhat-marketplace-tm7xt\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.032181 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-utilities\") pod \"redhat-marketplace-tm7xt\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.042377 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm7xt"] Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.138321 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k49rw\" (UniqueName: \"kubernetes.io/projected/b5752df4-d789-4fee-a67e-5329584c5120-kube-api-access-k49rw\") pod \"redhat-marketplace-tm7xt\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.138393 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-utilities\") pod \"redhat-marketplace-tm7xt\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.138432 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-catalog-content\") pod \"redhat-marketplace-tm7xt\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.138972 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-utilities\") pod \"redhat-marketplace-tm7xt\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.139033 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-catalog-content\") pod \"redhat-marketplace-tm7xt\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.167414 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k49rw\" (UniqueName: \"kubernetes.io/projected/b5752df4-d789-4fee-a67e-5329584c5120-kube-api-access-k49rw\") pod \"redhat-marketplace-tm7xt\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.361330 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:51:56 crc kubenswrapper[4945]: W1206 08:51:56.941040 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5752df4_d789_4fee_a67e_5329584c5120.slice/crio-27929112de1971cc9b421b09080132db3340d6c286f25c16744007a4bc5582fb WatchSource:0}: Error finding container 27929112de1971cc9b421b09080132db3340d6c286f25c16744007a4bc5582fb: Status 404 returned error can't find the container with id 27929112de1971cc9b421b09080132db3340d6c286f25c16744007a4bc5582fb Dec 06 08:51:56 crc kubenswrapper[4945]: I1206 08:51:56.943337 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm7xt"] Dec 06 08:51:57 crc kubenswrapper[4945]: I1206 08:51:57.430668 4945 generic.go:334] "Generic (PLEG): container finished" podID="b5752df4-d789-4fee-a67e-5329584c5120" containerID="309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99" exitCode=0 Dec 06 08:51:57 crc kubenswrapper[4945]: I1206 08:51:57.430720 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm7xt" event={"ID":"b5752df4-d789-4fee-a67e-5329584c5120","Type":"ContainerDied","Data":"309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99"} Dec 06 08:51:57 crc kubenswrapper[4945]: I1206 08:51:57.430962 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm7xt" event={"ID":"b5752df4-d789-4fee-a67e-5329584c5120","Type":"ContainerStarted","Data":"27929112de1971cc9b421b09080132db3340d6c286f25c16744007a4bc5582fb"} Dec 06 08:51:58 crc kubenswrapper[4945]: I1206 08:51:58.441572 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm7xt" event={"ID":"b5752df4-d789-4fee-a67e-5329584c5120","Type":"ContainerStarted","Data":"6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1"} Dec 06 08:51:59 crc kubenswrapper[4945]: I1206 08:51:59.451642 4945 generic.go:334] "Generic (PLEG): container finished" podID="b5752df4-d789-4fee-a67e-5329584c5120" containerID="6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1" exitCode=0 Dec 06 08:51:59 crc kubenswrapper[4945]: I1206 08:51:59.451702 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm7xt" event={"ID":"b5752df4-d789-4fee-a67e-5329584c5120","Type":"ContainerDied","Data":"6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1"} Dec 06 08:52:00 crc kubenswrapper[4945]: I1206 08:52:00.464555 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm7xt" event={"ID":"b5752df4-d789-4fee-a67e-5329584c5120","Type":"ContainerStarted","Data":"3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13"} Dec 06 08:52:00 crc kubenswrapper[4945]: I1206 08:52:00.495481 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tm7xt" podStartSLOduration=2.044376714 podStartE2EDuration="4.495459707s" podCreationTimestamp="2025-12-06 08:51:56 +0000 UTC" firstStartedPulling="2025-12-06 08:51:57.432336469 +0000 UTC m=+7130.887197513" lastFinishedPulling="2025-12-06 08:51:59.883419462 +0000 UTC m=+7133.338280506" observedRunningTime="2025-12-06 08:52:00.484083322 +0000 UTC m=+7133.938944366" watchObservedRunningTime="2025-12-06 08:52:00.495459707 +0000 UTC m=+7133.950320751" Dec 06 08:52:06 crc kubenswrapper[4945]: I1206 08:52:06.361941 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:52:06 crc kubenswrapper[4945]: I1206 08:52:06.362459 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:52:06 crc kubenswrapper[4945]: I1206 08:52:06.409687 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:52:06 crc kubenswrapper[4945]: I1206 08:52:06.560698 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:52:06 crc kubenswrapper[4945]: I1206 08:52:06.649043 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm7xt"] Dec 06 08:52:08 crc kubenswrapper[4945]: I1206 08:52:08.532383 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tm7xt" podUID="b5752df4-d789-4fee-a67e-5329584c5120" containerName="registry-server" containerID="cri-o://3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13" gracePeriod=2 Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.063212 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.207139 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-utilities\") pod \"b5752df4-d789-4fee-a67e-5329584c5120\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.207442 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-catalog-content\") pod \"b5752df4-d789-4fee-a67e-5329584c5120\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.207618 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k49rw\" (UniqueName: \"kubernetes.io/projected/b5752df4-d789-4fee-a67e-5329584c5120-kube-api-access-k49rw\") pod \"b5752df4-d789-4fee-a67e-5329584c5120\" (UID: \"b5752df4-d789-4fee-a67e-5329584c5120\") " Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.208099 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-utilities" (OuterVolumeSpecName: "utilities") pod "b5752df4-d789-4fee-a67e-5329584c5120" (UID: "b5752df4-d789-4fee-a67e-5329584c5120"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.208313 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.216028 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5752df4-d789-4fee-a67e-5329584c5120-kube-api-access-k49rw" (OuterVolumeSpecName: "kube-api-access-k49rw") pod "b5752df4-d789-4fee-a67e-5329584c5120" (UID: "b5752df4-d789-4fee-a67e-5329584c5120"). InnerVolumeSpecName "kube-api-access-k49rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.227201 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5752df4-d789-4fee-a67e-5329584c5120" (UID: "b5752df4-d789-4fee-a67e-5329584c5120"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.309888 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5752df4-d789-4fee-a67e-5329584c5120-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.309938 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k49rw\" (UniqueName: \"kubernetes.io/projected/b5752df4-d789-4fee-a67e-5329584c5120-kube-api-access-k49rw\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.544479 4945 generic.go:334] "Generic (PLEG): container finished" podID="b5752df4-d789-4fee-a67e-5329584c5120" containerID="3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13" exitCode=0 Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.544689 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm7xt" event={"ID":"b5752df4-d789-4fee-a67e-5329584c5120","Type":"ContainerDied","Data":"3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13"} Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.544856 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm7xt" event={"ID":"b5752df4-d789-4fee-a67e-5329584c5120","Type":"ContainerDied","Data":"27929112de1971cc9b421b09080132db3340d6c286f25c16744007a4bc5582fb"} Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.544880 4945 scope.go:117] "RemoveContainer" containerID="3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.544789 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm7xt" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.574555 4945 scope.go:117] "RemoveContainer" containerID="6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.589371 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm7xt"] Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.601415 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm7xt"] Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.619267 4945 scope.go:117] "RemoveContainer" containerID="309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.646774 4945 scope.go:117] "RemoveContainer" containerID="3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13" Dec 06 08:52:09 crc kubenswrapper[4945]: E1206 08:52:09.647968 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13\": container with ID starting with 3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13 not found: ID does not exist" containerID="3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.648005 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13"} err="failed to get container status \"3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13\": rpc error: code = NotFound desc = could not find container \"3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13\": container with ID starting with 3fa3e13c2b73a3a61a10db68c118323490d2c01723caa770f199be1fd8873e13 not found: ID does not exist" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.648026 4945 scope.go:117] "RemoveContainer" containerID="6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1" Dec 06 08:52:09 crc kubenswrapper[4945]: E1206 08:52:09.648990 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1\": container with ID starting with 6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1 not found: ID does not exist" containerID="6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.649021 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1"} err="failed to get container status \"6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1\": rpc error: code = NotFound desc = could not find container \"6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1\": container with ID starting with 6e417682711b6d98152492b01e3df0198d2ecfe71c0884784559ee895b07afb1 not found: ID does not exist" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.649036 4945 scope.go:117] "RemoveContainer" containerID="309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99" Dec 06 08:52:09 crc kubenswrapper[4945]: E1206 08:52:09.649388 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99\": container with ID starting with 309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99 not found: ID does not exist" containerID="309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99" Dec 06 08:52:09 crc kubenswrapper[4945]: I1206 08:52:09.649416 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99"} err="failed to get container status \"309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99\": rpc error: code = NotFound desc = could not find container \"309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99\": container with ID starting with 309944e8c1aca062106ccca68c1cf2a51288b6410ac466c7353cccc8bba11a99 not found: ID does not exist" Dec 06 08:52:10 crc kubenswrapper[4945]: I1206 08:52:10.964535 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5752df4-d789-4fee-a67e-5329584c5120" path="/var/lib/kubelet/pods/b5752df4-d789-4fee-a67e-5329584c5120/volumes" Dec 06 08:52:11 crc kubenswrapper[4945]: I1206 08:52:11.567360 4945 generic.go:334] "Generic (PLEG): container finished" podID="6fcc9e05-4e5d-434c-bd5b-382d452fc94c" containerID="36897258927ac269a46833c77a9d9c171d86738d060ef81041f0a3c5f4b7c0b6" exitCode=0 Dec 06 08:52:11 crc kubenswrapper[4945]: I1206 08:52:11.567426 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-gnccs" event={"ID":"6fcc9e05-4e5d-434c-bd5b-382d452fc94c","Type":"ContainerDied","Data":"36897258927ac269a46833c77a9d9c171d86738d060ef81041f0a3c5f4b7c0b6"} Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.074548 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.184957 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-dhcp-combined-ca-bundle\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185013 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ovn-combined-ca-bundle\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185113 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-libvirt-combined-ca-bundle\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185149 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-metadata-combined-ca-bundle\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185183 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ssh-key\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185235 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-telemetry-combined-ca-bundle\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185254 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-sriov-combined-ca-bundle\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185339 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-inventory\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185393 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-nova-combined-ca-bundle\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185458 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ceph\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185517 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-bootstrap-combined-ca-bundle\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.185541 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2np7j\" (UniqueName: \"kubernetes.io/projected/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-kube-api-access-2np7j\") pod \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\" (UID: \"6fcc9e05-4e5d-434c-bd5b-382d452fc94c\") " Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.191913 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ceph" (OuterVolumeSpecName: "ceph") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.191951 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-kube-api-access-2np7j" (OuterVolumeSpecName: "kube-api-access-2np7j") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "kube-api-access-2np7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.191960 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.191990 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.193258 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.193305 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.194454 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.195143 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.195485 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.216018 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.219992 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-inventory" (OuterVolumeSpecName: "inventory") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.231926 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6fcc9e05-4e5d-434c-bd5b-382d452fc94c" (UID: "6fcc9e05-4e5d-434c-bd5b-382d452fc94c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288394 4945 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288430 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288444 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288455 4945 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288466 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288474 4945 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288482 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2np7j\" (UniqueName: \"kubernetes.io/projected/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-kube-api-access-2np7j\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288494 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288503 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288514 4945 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288525 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.288534 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6fcc9e05-4e5d-434c-bd5b-382d452fc94c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.587375 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-gnccs" event={"ID":"6fcc9e05-4e5d-434c-bd5b-382d452fc94c","Type":"ContainerDied","Data":"28cf1370b46b43f6755383c2de4b76581c3d380bb41072ee07cfbc497c9f2f88"} Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.587415 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28cf1370b46b43f6755383c2de4b76581c3d380bb41072ee07cfbc497c9f2f88" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.587439 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-gnccs" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.670420 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-w27w6"] Dec 06 08:52:13 crc kubenswrapper[4945]: E1206 08:52:13.670882 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5752df4-d789-4fee-a67e-5329584c5120" containerName="registry-server" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.670903 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5752df4-d789-4fee-a67e-5329584c5120" containerName="registry-server" Dec 06 08:52:13 crc kubenswrapper[4945]: E1206 08:52:13.670925 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fcc9e05-4e5d-434c-bd5b-382d452fc94c" containerName="install-certs-openstack-openstack-cell1" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.670934 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fcc9e05-4e5d-434c-bd5b-382d452fc94c" containerName="install-certs-openstack-openstack-cell1" Dec 06 08:52:13 crc kubenswrapper[4945]: E1206 08:52:13.670956 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5752df4-d789-4fee-a67e-5329584c5120" containerName="extract-content" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.670963 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5752df4-d789-4fee-a67e-5329584c5120" containerName="extract-content" Dec 06 08:52:13 crc kubenswrapper[4945]: E1206 08:52:13.670971 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5752df4-d789-4fee-a67e-5329584c5120" containerName="extract-utilities" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.670977 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5752df4-d789-4fee-a67e-5329584c5120" containerName="extract-utilities" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.671176 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fcc9e05-4e5d-434c-bd5b-382d452fc94c" containerName="install-certs-openstack-openstack-cell1" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.671196 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5752df4-d789-4fee-a67e-5329584c5120" containerName="registry-server" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.672028 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.674272 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.674491 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.675777 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.678919 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.695986 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-inventory\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.696207 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.696253 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ceph\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.696471 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krjlt\" (UniqueName: \"kubernetes.io/projected/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-kube-api-access-krjlt\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.701078 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-w27w6"] Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.797867 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krjlt\" (UniqueName: \"kubernetes.io/projected/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-kube-api-access-krjlt\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.798033 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-inventory\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.798122 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.798183 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ceph\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.802662 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ceph\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.803771 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-inventory\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.804313 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.817204 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krjlt\" (UniqueName: \"kubernetes.io/projected/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-kube-api-access-krjlt\") pod \"ceph-client-openstack-openstack-cell1-w27w6\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:13 crc kubenswrapper[4945]: I1206 08:52:13.996761 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:14 crc kubenswrapper[4945]: I1206 08:52:14.528253 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-w27w6"] Dec 06 08:52:14 crc kubenswrapper[4945]: I1206 08:52:14.597964 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" event={"ID":"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2","Type":"ContainerStarted","Data":"d85cd1bdfa91e8ef37c60d45543b113914820209acfdac2b308d77fcb9846d9a"} Dec 06 08:52:15 crc kubenswrapper[4945]: I1206 08:52:15.607681 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" event={"ID":"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2","Type":"ContainerStarted","Data":"a2fe0b4e173b599e284a1e6b711509822c813cb9c1ec48e09bf880525921d3a5"} Dec 06 08:52:15 crc kubenswrapper[4945]: I1206 08:52:15.629270 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" podStartSLOduration=2.232428568 podStartE2EDuration="2.629250095s" podCreationTimestamp="2025-12-06 08:52:13 +0000 UTC" firstStartedPulling="2025-12-06 08:52:14.530890607 +0000 UTC m=+7147.985751651" lastFinishedPulling="2025-12-06 08:52:14.927712134 +0000 UTC m=+7148.382573178" observedRunningTime="2025-12-06 08:52:15.623375738 +0000 UTC m=+7149.078236802" watchObservedRunningTime="2025-12-06 08:52:15.629250095 +0000 UTC m=+7149.084111149" Dec 06 08:52:19 crc kubenswrapper[4945]: I1206 08:52:19.649265 4945 generic.go:334] "Generic (PLEG): container finished" podID="ac1cf84d-8334-415c-a6cc-3c65a26a7bb2" containerID="a2fe0b4e173b599e284a1e6b711509822c813cb9c1ec48e09bf880525921d3a5" exitCode=0 Dec 06 08:52:19 crc kubenswrapper[4945]: I1206 08:52:19.649377 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" event={"ID":"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2","Type":"ContainerDied","Data":"a2fe0b4e173b599e284a1e6b711509822c813cb9c1ec48e09bf880525921d3a5"} Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.086918 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.237241 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ssh-key\") pod \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.237379 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ceph\") pod \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.237473 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-inventory\") pod \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.237615 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krjlt\" (UniqueName: \"kubernetes.io/projected/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-kube-api-access-krjlt\") pod \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\" (UID: \"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2\") " Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.244490 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ceph" (OuterVolumeSpecName: "ceph") pod "ac1cf84d-8334-415c-a6cc-3c65a26a7bb2" (UID: "ac1cf84d-8334-415c-a6cc-3c65a26a7bb2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.244700 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-kube-api-access-krjlt" (OuterVolumeSpecName: "kube-api-access-krjlt") pod "ac1cf84d-8334-415c-a6cc-3c65a26a7bb2" (UID: "ac1cf84d-8334-415c-a6cc-3c65a26a7bb2"). InnerVolumeSpecName "kube-api-access-krjlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.266903 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-inventory" (OuterVolumeSpecName: "inventory") pod "ac1cf84d-8334-415c-a6cc-3c65a26a7bb2" (UID: "ac1cf84d-8334-415c-a6cc-3c65a26a7bb2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.273428 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ac1cf84d-8334-415c-a6cc-3c65a26a7bb2" (UID: "ac1cf84d-8334-415c-a6cc-3c65a26a7bb2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.340480 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.341187 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.341213 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krjlt\" (UniqueName: \"kubernetes.io/projected/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-kube-api-access-krjlt\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.341225 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac1cf84d-8334-415c-a6cc-3c65a26a7bb2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.670464 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" event={"ID":"ac1cf84d-8334-415c-a6cc-3c65a26a7bb2","Type":"ContainerDied","Data":"d85cd1bdfa91e8ef37c60d45543b113914820209acfdac2b308d77fcb9846d9a"} Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.670505 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d85cd1bdfa91e8ef37c60d45543b113914820209acfdac2b308d77fcb9846d9a" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.670555 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-w27w6" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.748216 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-j7cpj"] Dec 06 08:52:21 crc kubenswrapper[4945]: E1206 08:52:21.748754 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1cf84d-8334-415c-a6cc-3c65a26a7bb2" containerName="ceph-client-openstack-openstack-cell1" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.748772 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1cf84d-8334-415c-a6cc-3c65a26a7bb2" containerName="ceph-client-openstack-openstack-cell1" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.751769 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac1cf84d-8334-415c-a6cc-3c65a26a7bb2" containerName="ceph-client-openstack-openstack-cell1" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.752630 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.754694 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.754896 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.755150 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.755351 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.756335 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.763601 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-j7cpj"] Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.952572 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-inventory\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.955261 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ceph\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.955397 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.955490 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ssh-key\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.955833 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-452j2\" (UniqueName: \"kubernetes.io/projected/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-kube-api-access-452j2\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:21 crc kubenswrapper[4945]: I1206 08:52:21.956223 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.058944 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-inventory\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.059063 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ceph\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.059101 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.059152 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ssh-key\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.059258 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-452j2\" (UniqueName: \"kubernetes.io/projected/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-kube-api-access-452j2\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.059335 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.061067 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.065163 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ceph\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.065698 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-inventory\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.067632 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.071501 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ssh-key\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.082802 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-452j2\" (UniqueName: \"kubernetes.io/projected/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-kube-api-access-452j2\") pod \"ovn-openstack-openstack-cell1-j7cpj\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.375794 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:52:22 crc kubenswrapper[4945]: I1206 08:52:22.876556 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-j7cpj"] Dec 06 08:52:23 crc kubenswrapper[4945]: I1206 08:52:23.698414 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-j7cpj" event={"ID":"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02","Type":"ContainerStarted","Data":"63c1e654fa93bdfe4ab770438c1f9b977d97968c192d3d690351a7d54c872b3d"} Dec 06 08:52:23 crc kubenswrapper[4945]: I1206 08:52:23.698764 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-j7cpj" event={"ID":"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02","Type":"ContainerStarted","Data":"8d004deb84c41d7c4820dcb76aae596576fdd6fd8d88b031057b7f4c67b4b4ba"} Dec 06 08:52:23 crc kubenswrapper[4945]: I1206 08:52:23.725654 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-j7cpj" podStartSLOduration=2.326152138 podStartE2EDuration="2.725632077s" podCreationTimestamp="2025-12-06 08:52:21 +0000 UTC" firstStartedPulling="2025-12-06 08:52:22.88147528 +0000 UTC m=+7156.336336314" lastFinishedPulling="2025-12-06 08:52:23.280955209 +0000 UTC m=+7156.735816253" observedRunningTime="2025-12-06 08:52:23.71417496 +0000 UTC m=+7157.169036004" watchObservedRunningTime="2025-12-06 08:52:23.725632077 +0000 UTC m=+7157.180493121" Dec 06 08:52:38 crc kubenswrapper[4945]: I1206 08:52:38.795779 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:52:38 crc kubenswrapper[4945]: I1206 08:52:38.796416 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:53:08 crc kubenswrapper[4945]: I1206 08:53:08.796169 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:53:08 crc kubenswrapper[4945]: I1206 08:53:08.797452 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:53:25 crc kubenswrapper[4945]: I1206 08:53:25.267933 4945 generic.go:334] "Generic (PLEG): container finished" podID="4bbbd623-f02c-4b7b-ab9c-f33e570a7b02" containerID="63c1e654fa93bdfe4ab770438c1f9b977d97968c192d3d690351a7d54c872b3d" exitCode=0 Dec 06 08:53:25 crc kubenswrapper[4945]: I1206 08:53:25.268025 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-j7cpj" event={"ID":"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02","Type":"ContainerDied","Data":"63c1e654fa93bdfe4ab770438c1f9b977d97968c192d3d690351a7d54c872b3d"} Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.716786 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.756869 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovn-combined-ca-bundle\") pod \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.757342 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ssh-key\") pod \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.757376 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ceph\") pod \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.757406 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovncontroller-config-0\") pod \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.757450 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-inventory\") pod \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.757554 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-452j2\" (UniqueName: \"kubernetes.io/projected/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-kube-api-access-452j2\") pod \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\" (UID: \"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02\") " Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.768273 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ceph" (OuterVolumeSpecName: "ceph") pod "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02" (UID: "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.768322 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-kube-api-access-452j2" (OuterVolumeSpecName: "kube-api-access-452j2") pod "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02" (UID: "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02"). InnerVolumeSpecName "kube-api-access-452j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.768621 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02" (UID: "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.798304 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02" (UID: "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.800832 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02" (UID: "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.806872 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-inventory" (OuterVolumeSpecName: "inventory") pod "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02" (UID: "4bbbd623-f02c-4b7b-ab9c-f33e570a7b02"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.860809 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.860847 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.860856 4945 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.860869 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.860878 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-452j2\" (UniqueName: \"kubernetes.io/projected/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-kube-api-access-452j2\") on node \"crc\" DevicePath \"\"" Dec 06 08:53:26 crc kubenswrapper[4945]: I1206 08:53:26.860888 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbbd623-f02c-4b7b-ab9c-f33e570a7b02-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.288937 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-j7cpj" event={"ID":"4bbbd623-f02c-4b7b-ab9c-f33e570a7b02","Type":"ContainerDied","Data":"8d004deb84c41d7c4820dcb76aae596576fdd6fd8d88b031057b7f4c67b4b4ba"} Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.288980 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d004deb84c41d7c4820dcb76aae596576fdd6fd8d88b031057b7f4c67b4b4ba" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.289334 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-j7cpj" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.370853 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-5xr9h"] Dec 06 08:53:27 crc kubenswrapper[4945]: E1206 08:53:27.371438 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbbd623-f02c-4b7b-ab9c-f33e570a7b02" containerName="ovn-openstack-openstack-cell1" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.371462 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbbd623-f02c-4b7b-ab9c-f33e570a7b02" containerName="ovn-openstack-openstack-cell1" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.371722 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbbd623-f02c-4b7b-ab9c-f33e570a7b02" containerName="ovn-openstack-openstack-cell1" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.372582 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.375692 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.375917 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.375917 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.376038 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.378601 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.379729 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.394408 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-5xr9h"] Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.474048 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.474100 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.474131 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.474200 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.474301 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.474344 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8bxm\" (UniqueName: \"kubernetes.io/projected/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-kube-api-access-j8bxm\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.474599 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.576341 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.576433 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.576488 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.576560 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.576745 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.576999 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8bxm\" (UniqueName: \"kubernetes.io/projected/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-kube-api-access-j8bxm\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.577084 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.583139 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.586197 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.586893 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.589618 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.591559 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.591821 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.595249 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8bxm\" (UniqueName: \"kubernetes.io/projected/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-kube-api-access-j8bxm\") pod \"neutron-metadata-openstack-openstack-cell1-5xr9h\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:27 crc kubenswrapper[4945]: I1206 08:53:27.694788 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:53:28 crc kubenswrapper[4945]: I1206 08:53:28.267517 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-5xr9h"] Dec 06 08:53:28 crc kubenswrapper[4945]: I1206 08:53:28.299868 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" event={"ID":"8771b333-f1cc-4655-b6af-8e7eddcb2b7a","Type":"ContainerStarted","Data":"5965b146c51497bde5f9a30008d7ee2bbad16354d5be20bfc7f9a69f10b4f1c1"} Dec 06 08:53:29 crc kubenswrapper[4945]: I1206 08:53:29.330132 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" event={"ID":"8771b333-f1cc-4655-b6af-8e7eddcb2b7a","Type":"ContainerStarted","Data":"cc22d89c7d1d32d3350590f09a7a5b639726847b2f1f4a1f7a46a7c72960abb6"} Dec 06 08:53:29 crc kubenswrapper[4945]: I1206 08:53:29.362674 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" podStartSLOduration=1.956998761 podStartE2EDuration="2.362653046s" podCreationTimestamp="2025-12-06 08:53:27 +0000 UTC" firstStartedPulling="2025-12-06 08:53:28.271410487 +0000 UTC m=+7221.726271531" lastFinishedPulling="2025-12-06 08:53:28.677064772 +0000 UTC m=+7222.131925816" observedRunningTime="2025-12-06 08:53:29.353730437 +0000 UTC m=+7222.808591471" watchObservedRunningTime="2025-12-06 08:53:29.362653046 +0000 UTC m=+7222.817514090" Dec 06 08:53:38 crc kubenswrapper[4945]: I1206 08:53:38.795931 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:53:38 crc kubenswrapper[4945]: I1206 08:53:38.796731 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:53:38 crc kubenswrapper[4945]: I1206 08:53:38.796772 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 08:53:38 crc kubenswrapper[4945]: I1206 08:53:38.797426 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"77f6ccf1b6f2295e7b8cc956a5858de679eeb654affbd183de5ef1d7218d8eeb"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 08:53:38 crc kubenswrapper[4945]: I1206 08:53:38.797493 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://77f6ccf1b6f2295e7b8cc956a5858de679eeb654affbd183de5ef1d7218d8eeb" gracePeriod=600 Dec 06 08:53:39 crc kubenswrapper[4945]: I1206 08:53:39.438349 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="77f6ccf1b6f2295e7b8cc956a5858de679eeb654affbd183de5ef1d7218d8eeb" exitCode=0 Dec 06 08:53:39 crc kubenswrapper[4945]: I1206 08:53:39.438388 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"77f6ccf1b6f2295e7b8cc956a5858de679eeb654affbd183de5ef1d7218d8eeb"} Dec 06 08:53:39 crc kubenswrapper[4945]: I1206 08:53:39.439103 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774"} Dec 06 08:53:39 crc kubenswrapper[4945]: I1206 08:53:39.439130 4945 scope.go:117] "RemoveContainer" containerID="3fb6f421c542b24bb01ad6ffa7d12b6f6c513cbfa7a966b94a8a2d6601b72876" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.136603 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w4nmk"] Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.139476 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.151756 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w4nmk"] Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.321787 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-catalog-content\") pod \"community-operators-w4nmk\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.322143 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-utilities\") pod \"community-operators-w4nmk\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.322355 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p76tt\" (UniqueName: \"kubernetes.io/projected/83875b38-2500-4f95-b86d-8d92ece11823-kube-api-access-p76tt\") pod \"community-operators-w4nmk\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.425196 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-catalog-content\") pod \"community-operators-w4nmk\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.425650 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-utilities\") pod \"community-operators-w4nmk\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.425689 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p76tt\" (UniqueName: \"kubernetes.io/projected/83875b38-2500-4f95-b86d-8d92ece11823-kube-api-access-p76tt\") pod \"community-operators-w4nmk\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.426777 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-catalog-content\") pod \"community-operators-w4nmk\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.427050 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-utilities\") pod \"community-operators-w4nmk\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.447853 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p76tt\" (UniqueName: \"kubernetes.io/projected/83875b38-2500-4f95-b86d-8d92ece11823-kube-api-access-p76tt\") pod \"community-operators-w4nmk\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:09 crc kubenswrapper[4945]: I1206 08:54:09.459195 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:10 crc kubenswrapper[4945]: I1206 08:54:10.082234 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w4nmk"] Dec 06 08:54:10 crc kubenswrapper[4945]: I1206 08:54:10.730652 4945 generic.go:334] "Generic (PLEG): container finished" podID="83875b38-2500-4f95-b86d-8d92ece11823" containerID="61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95" exitCode=0 Dec 06 08:54:10 crc kubenswrapper[4945]: I1206 08:54:10.730722 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4nmk" event={"ID":"83875b38-2500-4f95-b86d-8d92ece11823","Type":"ContainerDied","Data":"61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95"} Dec 06 08:54:10 crc kubenswrapper[4945]: I1206 08:54:10.730979 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4nmk" event={"ID":"83875b38-2500-4f95-b86d-8d92ece11823","Type":"ContainerStarted","Data":"508645cd9cad11fafd51bcf1bbbe8b468a64cf371ee726b6a2671128f1232868"} Dec 06 08:54:11 crc kubenswrapper[4945]: I1206 08:54:11.741983 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4nmk" event={"ID":"83875b38-2500-4f95-b86d-8d92ece11823","Type":"ContainerStarted","Data":"7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e"} Dec 06 08:54:12 crc kubenswrapper[4945]: I1206 08:54:12.754184 4945 generic.go:334] "Generic (PLEG): container finished" podID="83875b38-2500-4f95-b86d-8d92ece11823" containerID="7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e" exitCode=0 Dec 06 08:54:12 crc kubenswrapper[4945]: I1206 08:54:12.754361 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4nmk" event={"ID":"83875b38-2500-4f95-b86d-8d92ece11823","Type":"ContainerDied","Data":"7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e"} Dec 06 08:54:13 crc kubenswrapper[4945]: I1206 08:54:13.766333 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4nmk" event={"ID":"83875b38-2500-4f95-b86d-8d92ece11823","Type":"ContainerStarted","Data":"50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab"} Dec 06 08:54:13 crc kubenswrapper[4945]: I1206 08:54:13.783227 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w4nmk" podStartSLOduration=2.275752105 podStartE2EDuration="4.783208446s" podCreationTimestamp="2025-12-06 08:54:09 +0000 UTC" firstStartedPulling="2025-12-06 08:54:10.732761826 +0000 UTC m=+7264.187622870" lastFinishedPulling="2025-12-06 08:54:13.240218167 +0000 UTC m=+7266.695079211" observedRunningTime="2025-12-06 08:54:13.782734333 +0000 UTC m=+7267.237595377" watchObservedRunningTime="2025-12-06 08:54:13.783208446 +0000 UTC m=+7267.238069490" Dec 06 08:54:19 crc kubenswrapper[4945]: I1206 08:54:19.459575 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:19 crc kubenswrapper[4945]: I1206 08:54:19.460113 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:19 crc kubenswrapper[4945]: I1206 08:54:19.505328 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:19 crc kubenswrapper[4945]: I1206 08:54:19.874894 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:19 crc kubenswrapper[4945]: I1206 08:54:19.920095 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w4nmk"] Dec 06 08:54:21 crc kubenswrapper[4945]: I1206 08:54:21.835126 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w4nmk" podUID="83875b38-2500-4f95-b86d-8d92ece11823" containerName="registry-server" containerID="cri-o://50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab" gracePeriod=2 Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.823169 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.837835 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-catalog-content\") pod \"83875b38-2500-4f95-b86d-8d92ece11823\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.837906 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p76tt\" (UniqueName: \"kubernetes.io/projected/83875b38-2500-4f95-b86d-8d92ece11823-kube-api-access-p76tt\") pod \"83875b38-2500-4f95-b86d-8d92ece11823\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.838029 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-utilities\") pod \"83875b38-2500-4f95-b86d-8d92ece11823\" (UID: \"83875b38-2500-4f95-b86d-8d92ece11823\") " Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.839020 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-utilities" (OuterVolumeSpecName: "utilities") pod "83875b38-2500-4f95-b86d-8d92ece11823" (UID: "83875b38-2500-4f95-b86d-8d92ece11823"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.885627 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83875b38-2500-4f95-b86d-8d92ece11823-kube-api-access-p76tt" (OuterVolumeSpecName: "kube-api-access-p76tt") pod "83875b38-2500-4f95-b86d-8d92ece11823" (UID: "83875b38-2500-4f95-b86d-8d92ece11823"). InnerVolumeSpecName "kube-api-access-p76tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.898973 4945 generic.go:334] "Generic (PLEG): container finished" podID="83875b38-2500-4f95-b86d-8d92ece11823" containerID="50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab" exitCode=0 Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.899038 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4nmk" event={"ID":"83875b38-2500-4f95-b86d-8d92ece11823","Type":"ContainerDied","Data":"50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab"} Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.899067 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4nmk" event={"ID":"83875b38-2500-4f95-b86d-8d92ece11823","Type":"ContainerDied","Data":"508645cd9cad11fafd51bcf1bbbe8b468a64cf371ee726b6a2671128f1232868"} Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.899084 4945 scope.go:117] "RemoveContainer" containerID="50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab" Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.899215 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4nmk" Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.906010 4945 generic.go:334] "Generic (PLEG): container finished" podID="8771b333-f1cc-4655-b6af-8e7eddcb2b7a" containerID="cc22d89c7d1d32d3350590f09a7a5b639726847b2f1f4a1f7a46a7c72960abb6" exitCode=0 Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.906054 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" event={"ID":"8771b333-f1cc-4655-b6af-8e7eddcb2b7a","Type":"ContainerDied","Data":"cc22d89c7d1d32d3350590f09a7a5b639726847b2f1f4a1f7a46a7c72960abb6"} Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.907812 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83875b38-2500-4f95-b86d-8d92ece11823" (UID: "83875b38-2500-4f95-b86d-8d92ece11823"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.929185 4945 scope.go:117] "RemoveContainer" containerID="7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e" Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.941667 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.941701 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p76tt\" (UniqueName: \"kubernetes.io/projected/83875b38-2500-4f95-b86d-8d92ece11823-kube-api-access-p76tt\") on node \"crc\" DevicePath \"\"" Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.941712 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83875b38-2500-4f95-b86d-8d92ece11823-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:54:22 crc kubenswrapper[4945]: I1206 08:54:22.962143 4945 scope.go:117] "RemoveContainer" containerID="61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95" Dec 06 08:54:23 crc kubenswrapper[4945]: I1206 08:54:23.006433 4945 scope.go:117] "RemoveContainer" containerID="50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab" Dec 06 08:54:23 crc kubenswrapper[4945]: E1206 08:54:23.006905 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab\": container with ID starting with 50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab not found: ID does not exist" containerID="50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab" Dec 06 08:54:23 crc kubenswrapper[4945]: I1206 08:54:23.006951 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab"} err="failed to get container status \"50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab\": rpc error: code = NotFound desc = could not find container \"50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab\": container with ID starting with 50565b5ecadc180e19d1dae5a27191c379ba34a357f798e48a4bcb36b9b2d2ab not found: ID does not exist" Dec 06 08:54:23 crc kubenswrapper[4945]: I1206 08:54:23.006995 4945 scope.go:117] "RemoveContainer" containerID="7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e" Dec 06 08:54:23 crc kubenswrapper[4945]: E1206 08:54:23.007385 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e\": container with ID starting with 7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e not found: ID does not exist" containerID="7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e" Dec 06 08:54:23 crc kubenswrapper[4945]: I1206 08:54:23.007435 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e"} err="failed to get container status \"7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e\": rpc error: code = NotFound desc = could not find container \"7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e\": container with ID starting with 7263bacfbd1699a9641fdd29e75abd2920c50f7d56782e0086ad85597bbcee3e not found: ID does not exist" Dec 06 08:54:23 crc kubenswrapper[4945]: I1206 08:54:23.007465 4945 scope.go:117] "RemoveContainer" containerID="61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95" Dec 06 08:54:23 crc kubenswrapper[4945]: E1206 08:54:23.007880 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95\": container with ID starting with 61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95 not found: ID does not exist" containerID="61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95" Dec 06 08:54:23 crc kubenswrapper[4945]: I1206 08:54:23.007943 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95"} err="failed to get container status \"61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95\": rpc error: code = NotFound desc = could not find container \"61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95\": container with ID starting with 61d9405c32412622e717a0b5466946a747c1b84891a752bca6ab7e02c3bfee95 not found: ID does not exist" Dec 06 08:54:23 crc kubenswrapper[4945]: I1206 08:54:23.227150 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w4nmk"] Dec 06 08:54:23 crc kubenswrapper[4945]: I1206 08:54:23.236123 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w4nmk"] Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.438636 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.473115 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ceph\") pod \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.473219 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-inventory\") pod \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.473290 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8bxm\" (UniqueName: \"kubernetes.io/projected/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-kube-api-access-j8bxm\") pod \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.473334 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ssh-key\") pod \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.473475 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-metadata-combined-ca-bundle\") pod \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.473553 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.473647 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-nova-metadata-neutron-config-0\") pod \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\" (UID: \"8771b333-f1cc-4655-b6af-8e7eddcb2b7a\") " Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.482623 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-kube-api-access-j8bxm" (OuterVolumeSpecName: "kube-api-access-j8bxm") pod "8771b333-f1cc-4655-b6af-8e7eddcb2b7a" (UID: "8771b333-f1cc-4655-b6af-8e7eddcb2b7a"). InnerVolumeSpecName "kube-api-access-j8bxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.484050 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ceph" (OuterVolumeSpecName: "ceph") pod "8771b333-f1cc-4655-b6af-8e7eddcb2b7a" (UID: "8771b333-f1cc-4655-b6af-8e7eddcb2b7a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.484245 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "8771b333-f1cc-4655-b6af-8e7eddcb2b7a" (UID: "8771b333-f1cc-4655-b6af-8e7eddcb2b7a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.507475 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-inventory" (OuterVolumeSpecName: "inventory") pod "8771b333-f1cc-4655-b6af-8e7eddcb2b7a" (UID: "8771b333-f1cc-4655-b6af-8e7eddcb2b7a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.507831 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "8771b333-f1cc-4655-b6af-8e7eddcb2b7a" (UID: "8771b333-f1cc-4655-b6af-8e7eddcb2b7a"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.511415 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8771b333-f1cc-4655-b6af-8e7eddcb2b7a" (UID: "8771b333-f1cc-4655-b6af-8e7eddcb2b7a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.520527 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "8771b333-f1cc-4655-b6af-8e7eddcb2b7a" (UID: "8771b333-f1cc-4655-b6af-8e7eddcb2b7a"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.575689 4945 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.575724 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.575740 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.575755 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8bxm\" (UniqueName: \"kubernetes.io/projected/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-kube-api-access-j8bxm\") on node \"crc\" DevicePath \"\"" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.575767 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.575778 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.575791 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8771b333-f1cc-4655-b6af-8e7eddcb2b7a-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.928947 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" event={"ID":"8771b333-f1cc-4655-b6af-8e7eddcb2b7a","Type":"ContainerDied","Data":"5965b146c51497bde5f9a30008d7ee2bbad16354d5be20bfc7f9a69f10b4f1c1"} Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.929263 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5965b146c51497bde5f9a30008d7ee2bbad16354d5be20bfc7f9a69f10b4f1c1" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.929024 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-5xr9h" Dec 06 08:54:24 crc kubenswrapper[4945]: I1206 08:54:24.973470 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83875b38-2500-4f95-b86d-8d92ece11823" path="/var/lib/kubelet/pods/83875b38-2500-4f95-b86d-8d92ece11823/volumes" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.027405 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-vh6xt"] Dec 06 08:54:25 crc kubenswrapper[4945]: E1206 08:54:25.027902 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8771b333-f1cc-4655-b6af-8e7eddcb2b7a" containerName="neutron-metadata-openstack-openstack-cell1" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.027921 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8771b333-f1cc-4655-b6af-8e7eddcb2b7a" containerName="neutron-metadata-openstack-openstack-cell1" Dec 06 08:54:25 crc kubenswrapper[4945]: E1206 08:54:25.027956 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83875b38-2500-4f95-b86d-8d92ece11823" containerName="extract-utilities" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.027964 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="83875b38-2500-4f95-b86d-8d92ece11823" containerName="extract-utilities" Dec 06 08:54:25 crc kubenswrapper[4945]: E1206 08:54:25.027973 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83875b38-2500-4f95-b86d-8d92ece11823" containerName="extract-content" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.027979 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="83875b38-2500-4f95-b86d-8d92ece11823" containerName="extract-content" Dec 06 08:54:25 crc kubenswrapper[4945]: E1206 08:54:25.027990 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83875b38-2500-4f95-b86d-8d92ece11823" containerName="registry-server" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.027996 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="83875b38-2500-4f95-b86d-8d92ece11823" containerName="registry-server" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.028208 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8771b333-f1cc-4655-b6af-8e7eddcb2b7a" containerName="neutron-metadata-openstack-openstack-cell1" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.028241 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="83875b38-2500-4f95-b86d-8d92ece11823" containerName="registry-server" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.029020 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.031615 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.031781 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.031893 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.032919 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.033958 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-vh6xt"] Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.034893 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.093357 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.093441 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.093481 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ssh-key\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.093537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-inventory\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.093557 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nbn6\" (UniqueName: \"kubernetes.io/projected/c1bde47d-052e-4406-86c7-9cf9c8cdf294-kube-api-access-5nbn6\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.093576 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ceph\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.196032 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.196124 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.196171 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ssh-key\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.196239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-inventory\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.196265 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nbn6\" (UniqueName: \"kubernetes.io/projected/c1bde47d-052e-4406-86c7-9cf9c8cdf294-kube-api-access-5nbn6\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.196306 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ceph\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.199883 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ceph\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.200522 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.200877 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.201059 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-inventory\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.201215 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ssh-key\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.216198 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nbn6\" (UniqueName: \"kubernetes.io/projected/c1bde47d-052e-4406-86c7-9cf9c8cdf294-kube-api-access-5nbn6\") pod \"libvirt-openstack-openstack-cell1-vh6xt\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.387884 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.914678 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-vh6xt"] Dec 06 08:54:25 crc kubenswrapper[4945]: W1206 08:54:25.917629 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1bde47d_052e_4406_86c7_9cf9c8cdf294.slice/crio-7398c008abebf0fad6a40fc8e9e200d7b814f32b03f5ea8d1c07d9ea81320474 WatchSource:0}: Error finding container 7398c008abebf0fad6a40fc8e9e200d7b814f32b03f5ea8d1c07d9ea81320474: Status 404 returned error can't find the container with id 7398c008abebf0fad6a40fc8e9e200d7b814f32b03f5ea8d1c07d9ea81320474 Dec 06 08:54:25 crc kubenswrapper[4945]: I1206 08:54:25.939898 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" event={"ID":"c1bde47d-052e-4406-86c7-9cf9c8cdf294","Type":"ContainerStarted","Data":"7398c008abebf0fad6a40fc8e9e200d7b814f32b03f5ea8d1c07d9ea81320474"} Dec 06 08:54:26 crc kubenswrapper[4945]: I1206 08:54:26.949433 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" event={"ID":"c1bde47d-052e-4406-86c7-9cf9c8cdf294","Type":"ContainerStarted","Data":"e8e9403dfc2182378df2009862c7d9887baba14dae2e23ebad42bd5cf8967b72"} Dec 06 08:54:26 crc kubenswrapper[4945]: I1206 08:54:26.997496 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" podStartSLOduration=2.468906411 podStartE2EDuration="2.997476634s" podCreationTimestamp="2025-12-06 08:54:24 +0000 UTC" firstStartedPulling="2025-12-06 08:54:25.919627294 +0000 UTC m=+7279.374488338" lastFinishedPulling="2025-12-06 08:54:26.448197527 +0000 UTC m=+7279.903058561" observedRunningTime="2025-12-06 08:54:26.971789197 +0000 UTC m=+7280.426650241" watchObservedRunningTime="2025-12-06 08:54:26.997476634 +0000 UTC m=+7280.452337678" Dec 06 08:56:08 crc kubenswrapper[4945]: I1206 08:56:08.796478 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:56:08 crc kubenswrapper[4945]: I1206 08:56:08.797852 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:56:38 crc kubenswrapper[4945]: I1206 08:56:38.796206 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:56:38 crc kubenswrapper[4945]: I1206 08:56:38.796870 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:57:08 crc kubenswrapper[4945]: I1206 08:57:08.795844 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 08:57:08 crc kubenswrapper[4945]: I1206 08:57:08.796511 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 08:57:08 crc kubenswrapper[4945]: I1206 08:57:08.796569 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 08:57:08 crc kubenswrapper[4945]: I1206 08:57:08.797417 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 08:57:08 crc kubenswrapper[4945]: I1206 08:57:08.797465 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" gracePeriod=600 Dec 06 08:57:08 crc kubenswrapper[4945]: E1206 08:57:08.921162 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:57:09 crc kubenswrapper[4945]: I1206 08:57:09.480683 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" exitCode=0 Dec 06 08:57:09 crc kubenswrapper[4945]: I1206 08:57:09.480739 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774"} Dec 06 08:57:09 crc kubenswrapper[4945]: I1206 08:57:09.480815 4945 scope.go:117] "RemoveContainer" containerID="77f6ccf1b6f2295e7b8cc956a5858de679eeb654affbd183de5ef1d7218d8eeb" Dec 06 08:57:09 crc kubenswrapper[4945]: I1206 08:57:09.481510 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:57:09 crc kubenswrapper[4945]: E1206 08:57:09.481939 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:57:19 crc kubenswrapper[4945]: I1206 08:57:19.953134 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:57:19 crc kubenswrapper[4945]: E1206 08:57:19.953907 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:57:22 crc kubenswrapper[4945]: I1206 08:57:22.906762 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fffbq"] Dec 06 08:57:22 crc kubenswrapper[4945]: I1206 08:57:22.910191 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:22 crc kubenswrapper[4945]: I1206 08:57:22.915447 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fffbq"] Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.039110 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-utilities\") pod \"redhat-operators-fffbq\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.039349 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-catalog-content\") pod \"redhat-operators-fffbq\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.039592 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hvkg\" (UniqueName: \"kubernetes.io/projected/2c9088ba-430d-40a7-8d88-84abe9901b46-kube-api-access-6hvkg\") pod \"redhat-operators-fffbq\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.142019 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hvkg\" (UniqueName: \"kubernetes.io/projected/2c9088ba-430d-40a7-8d88-84abe9901b46-kube-api-access-6hvkg\") pod \"redhat-operators-fffbq\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.142089 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-utilities\") pod \"redhat-operators-fffbq\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.142211 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-catalog-content\") pod \"redhat-operators-fffbq\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.142757 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-utilities\") pod \"redhat-operators-fffbq\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.142774 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-catalog-content\") pod \"redhat-operators-fffbq\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.164148 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hvkg\" (UniqueName: \"kubernetes.io/projected/2c9088ba-430d-40a7-8d88-84abe9901b46-kube-api-access-6hvkg\") pod \"redhat-operators-fffbq\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.237084 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:23 crc kubenswrapper[4945]: I1206 08:57:23.700012 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fffbq"] Dec 06 08:57:24 crc kubenswrapper[4945]: I1206 08:57:24.620413 4945 generic.go:334] "Generic (PLEG): container finished" podID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerID="029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6" exitCode=0 Dec 06 08:57:24 crc kubenswrapper[4945]: I1206 08:57:24.620516 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fffbq" event={"ID":"2c9088ba-430d-40a7-8d88-84abe9901b46","Type":"ContainerDied","Data":"029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6"} Dec 06 08:57:24 crc kubenswrapper[4945]: I1206 08:57:24.620845 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fffbq" event={"ID":"2c9088ba-430d-40a7-8d88-84abe9901b46","Type":"ContainerStarted","Data":"92348ab06a742f90edb8a00001ca24546426c66ece96eb7fbf64b7e90fe805dc"} Dec 06 08:57:24 crc kubenswrapper[4945]: I1206 08:57:24.622800 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 08:57:26 crc kubenswrapper[4945]: I1206 08:57:26.641937 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fffbq" event={"ID":"2c9088ba-430d-40a7-8d88-84abe9901b46","Type":"ContainerStarted","Data":"e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143"} Dec 06 08:57:29 crc kubenswrapper[4945]: I1206 08:57:29.671597 4945 generic.go:334] "Generic (PLEG): container finished" podID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerID="e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143" exitCode=0 Dec 06 08:57:29 crc kubenswrapper[4945]: I1206 08:57:29.671645 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fffbq" event={"ID":"2c9088ba-430d-40a7-8d88-84abe9901b46","Type":"ContainerDied","Data":"e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143"} Dec 06 08:57:30 crc kubenswrapper[4945]: I1206 08:57:30.683667 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fffbq" event={"ID":"2c9088ba-430d-40a7-8d88-84abe9901b46","Type":"ContainerStarted","Data":"9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3"} Dec 06 08:57:30 crc kubenswrapper[4945]: I1206 08:57:30.713195 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fffbq" podStartSLOduration=2.985396378 podStartE2EDuration="8.713169952s" podCreationTimestamp="2025-12-06 08:57:22 +0000 UTC" firstStartedPulling="2025-12-06 08:57:24.622570016 +0000 UTC m=+7458.077431060" lastFinishedPulling="2025-12-06 08:57:30.35034358 +0000 UTC m=+7463.805204634" observedRunningTime="2025-12-06 08:57:30.708810665 +0000 UTC m=+7464.163671709" watchObservedRunningTime="2025-12-06 08:57:30.713169952 +0000 UTC m=+7464.168030986" Dec 06 08:57:33 crc kubenswrapper[4945]: I1206 08:57:33.237500 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:33 crc kubenswrapper[4945]: I1206 08:57:33.237792 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:33 crc kubenswrapper[4945]: I1206 08:57:33.953670 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:57:33 crc kubenswrapper[4945]: E1206 08:57:33.954294 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:57:34 crc kubenswrapper[4945]: I1206 08:57:34.286010 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fffbq" podUID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerName="registry-server" probeResult="failure" output=< Dec 06 08:57:34 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 08:57:34 crc kubenswrapper[4945]: > Dec 06 08:57:43 crc kubenswrapper[4945]: I1206 08:57:43.283802 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:43 crc kubenswrapper[4945]: I1206 08:57:43.333522 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:43 crc kubenswrapper[4945]: I1206 08:57:43.521079 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fffbq"] Dec 06 08:57:44 crc kubenswrapper[4945]: I1206 08:57:44.819689 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fffbq" podUID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerName="registry-server" containerID="cri-o://9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3" gracePeriod=2 Dec 06 08:57:44 crc kubenswrapper[4945]: I1206 08:57:44.954073 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:57:44 crc kubenswrapper[4945]: E1206 08:57:44.954724 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.330731 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.427557 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-utilities\") pod \"2c9088ba-430d-40a7-8d88-84abe9901b46\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.427631 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hvkg\" (UniqueName: \"kubernetes.io/projected/2c9088ba-430d-40a7-8d88-84abe9901b46-kube-api-access-6hvkg\") pod \"2c9088ba-430d-40a7-8d88-84abe9901b46\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.427807 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-catalog-content\") pod \"2c9088ba-430d-40a7-8d88-84abe9901b46\" (UID: \"2c9088ba-430d-40a7-8d88-84abe9901b46\") " Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.429419 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-utilities" (OuterVolumeSpecName: "utilities") pod "2c9088ba-430d-40a7-8d88-84abe9901b46" (UID: "2c9088ba-430d-40a7-8d88-84abe9901b46"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.433910 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c9088ba-430d-40a7-8d88-84abe9901b46-kube-api-access-6hvkg" (OuterVolumeSpecName: "kube-api-access-6hvkg") pod "2c9088ba-430d-40a7-8d88-84abe9901b46" (UID: "2c9088ba-430d-40a7-8d88-84abe9901b46"). InnerVolumeSpecName "kube-api-access-6hvkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.531002 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.531693 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hvkg\" (UniqueName: \"kubernetes.io/projected/2c9088ba-430d-40a7-8d88-84abe9901b46-kube-api-access-6hvkg\") on node \"crc\" DevicePath \"\"" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.545053 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c9088ba-430d-40a7-8d88-84abe9901b46" (UID: "2c9088ba-430d-40a7-8d88-84abe9901b46"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.633696 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9088ba-430d-40a7-8d88-84abe9901b46-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.831402 4945 generic.go:334] "Generic (PLEG): container finished" podID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerID="9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3" exitCode=0 Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.831483 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fffbq" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.831472 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fffbq" event={"ID":"2c9088ba-430d-40a7-8d88-84abe9901b46","Type":"ContainerDied","Data":"9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3"} Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.831949 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fffbq" event={"ID":"2c9088ba-430d-40a7-8d88-84abe9901b46","Type":"ContainerDied","Data":"92348ab06a742f90edb8a00001ca24546426c66ece96eb7fbf64b7e90fe805dc"} Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.831997 4945 scope.go:117] "RemoveContainer" containerID="9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.866980 4945 scope.go:117] "RemoveContainer" containerID="e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.885855 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fffbq"] Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.895749 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fffbq"] Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.900106 4945 scope.go:117] "RemoveContainer" containerID="029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.946933 4945 scope.go:117] "RemoveContainer" containerID="9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3" Dec 06 08:57:45 crc kubenswrapper[4945]: E1206 08:57:45.947396 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3\": container with ID starting with 9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3 not found: ID does not exist" containerID="9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.947451 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3"} err="failed to get container status \"9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3\": rpc error: code = NotFound desc = could not find container \"9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3\": container with ID starting with 9ad8b75c09c9ed7f38dad9f9099fd8df983d4e6e37c89aee80f50766580e81b3 not found: ID does not exist" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.947491 4945 scope.go:117] "RemoveContainer" containerID="e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143" Dec 06 08:57:45 crc kubenswrapper[4945]: E1206 08:57:45.948047 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143\": container with ID starting with e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143 not found: ID does not exist" containerID="e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.948070 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143"} err="failed to get container status \"e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143\": rpc error: code = NotFound desc = could not find container \"e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143\": container with ID starting with e67839cd3da824ef6005e386a071a95af05c632abe8fbdaec00b66735434c143 not found: ID does not exist" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.948084 4945 scope.go:117] "RemoveContainer" containerID="029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6" Dec 06 08:57:45 crc kubenswrapper[4945]: E1206 08:57:45.948274 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6\": container with ID starting with 029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6 not found: ID does not exist" containerID="029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6" Dec 06 08:57:45 crc kubenswrapper[4945]: I1206 08:57:45.948335 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6"} err="failed to get container status \"029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6\": rpc error: code = NotFound desc = could not find container \"029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6\": container with ID starting with 029b85689f016382e83107e79243dca3c59b576593f325cc375cdd96d87dcbe6 not found: ID does not exist" Dec 06 08:57:46 crc kubenswrapper[4945]: I1206 08:57:46.966769 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c9088ba-430d-40a7-8d88-84abe9901b46" path="/var/lib/kubelet/pods/2c9088ba-430d-40a7-8d88-84abe9901b46/volumes" Dec 06 08:57:56 crc kubenswrapper[4945]: I1206 08:57:56.959756 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:57:56 crc kubenswrapper[4945]: E1206 08:57:56.960513 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:58:10 crc kubenswrapper[4945]: I1206 08:58:10.953806 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:58:10 crc kubenswrapper[4945]: E1206 08:58:10.954664 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:58:24 crc kubenswrapper[4945]: I1206 08:58:24.953789 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:58:24 crc kubenswrapper[4945]: E1206 08:58:24.954737 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:58:37 crc kubenswrapper[4945]: I1206 08:58:37.953749 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:58:37 crc kubenswrapper[4945]: E1206 08:58:37.954639 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:58:51 crc kubenswrapper[4945]: I1206 08:58:51.953485 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:58:51 crc kubenswrapper[4945]: E1206 08:58:51.955423 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:59:02 crc kubenswrapper[4945]: I1206 08:59:02.954571 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:59:02 crc kubenswrapper[4945]: E1206 08:59:02.955450 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:59:07 crc kubenswrapper[4945]: I1206 08:59:07.601309 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" event={"ID":"c1bde47d-052e-4406-86c7-9cf9c8cdf294","Type":"ContainerDied","Data":"e8e9403dfc2182378df2009862c7d9887baba14dae2e23ebad42bd5cf8967b72"} Dec 06 08:59:07 crc kubenswrapper[4945]: I1206 08:59:07.601266 4945 generic.go:334] "Generic (PLEG): container finished" podID="c1bde47d-052e-4406-86c7-9cf9c8cdf294" containerID="e8e9403dfc2182378df2009862c7d9887baba14dae2e23ebad42bd5cf8967b72" exitCode=0 Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.076938 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.222224 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ssh-key\") pod \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.222508 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nbn6\" (UniqueName: \"kubernetes.io/projected/c1bde47d-052e-4406-86c7-9cf9c8cdf294-kube-api-access-5nbn6\") pod \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.222649 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-secret-0\") pod \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.222803 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ceph\") pod \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.222862 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-combined-ca-bundle\") pod \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.222898 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-inventory\") pod \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\" (UID: \"c1bde47d-052e-4406-86c7-9cf9c8cdf294\") " Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.228927 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1bde47d-052e-4406-86c7-9cf9c8cdf294-kube-api-access-5nbn6" (OuterVolumeSpecName: "kube-api-access-5nbn6") pod "c1bde47d-052e-4406-86c7-9cf9c8cdf294" (UID: "c1bde47d-052e-4406-86c7-9cf9c8cdf294"). InnerVolumeSpecName "kube-api-access-5nbn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.228948 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c1bde47d-052e-4406-86c7-9cf9c8cdf294" (UID: "c1bde47d-052e-4406-86c7-9cf9c8cdf294"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.228881 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ceph" (OuterVolumeSpecName: "ceph") pod "c1bde47d-052e-4406-86c7-9cf9c8cdf294" (UID: "c1bde47d-052e-4406-86c7-9cf9c8cdf294"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.252993 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "c1bde47d-052e-4406-86c7-9cf9c8cdf294" (UID: "c1bde47d-052e-4406-86c7-9cf9c8cdf294"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.258226 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c1bde47d-052e-4406-86c7-9cf9c8cdf294" (UID: "c1bde47d-052e-4406-86c7-9cf9c8cdf294"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.266642 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-inventory" (OuterVolumeSpecName: "inventory") pod "c1bde47d-052e-4406-86c7-9cf9c8cdf294" (UID: "c1bde47d-052e-4406-86c7-9cf9c8cdf294"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.325981 4945 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.326265 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.327563 4945 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.327589 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.327605 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1bde47d-052e-4406-86c7-9cf9c8cdf294-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.327618 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nbn6\" (UniqueName: \"kubernetes.io/projected/c1bde47d-052e-4406-86c7-9cf9c8cdf294-kube-api-access-5nbn6\") on node \"crc\" DevicePath \"\"" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.621169 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" event={"ID":"c1bde47d-052e-4406-86c7-9cf9c8cdf294","Type":"ContainerDied","Data":"7398c008abebf0fad6a40fc8e9e200d7b814f32b03f5ea8d1c07d9ea81320474"} Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.621209 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7398c008abebf0fad6a40fc8e9e200d7b814f32b03f5ea8d1c07d9ea81320474" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.621261 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-vh6xt" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.745986 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-4pkmq"] Dec 06 08:59:09 crc kubenswrapper[4945]: E1206 08:59:09.746572 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerName="registry-server" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.746587 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerName="registry-server" Dec 06 08:59:09 crc kubenswrapper[4945]: E1206 08:59:09.746622 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerName="extract-utilities" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.746630 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerName="extract-utilities" Dec 06 08:59:09 crc kubenswrapper[4945]: E1206 08:59:09.746651 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerName="extract-content" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.746659 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerName="extract-content" Dec 06 08:59:09 crc kubenswrapper[4945]: E1206 08:59:09.746674 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1bde47d-052e-4406-86c7-9cf9c8cdf294" containerName="libvirt-openstack-openstack-cell1" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.746686 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1bde47d-052e-4406-86c7-9cf9c8cdf294" containerName="libvirt-openstack-openstack-cell1" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.746901 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1bde47d-052e-4406-86c7-9cf9c8cdf294" containerName="libvirt-openstack-openstack-cell1" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.746933 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c9088ba-430d-40a7-8d88-84abe9901b46" containerName="registry-server" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.747900 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.751510 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.751745 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.751882 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.752340 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.752467 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.752507 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.752770 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.767812 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-4pkmq"] Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.900129 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.900522 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.900654 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-inventory\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.900802 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ceph\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.900913 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvqq6\" (UniqueName: \"kubernetes.io/projected/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-kube-api-access-nvqq6\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.901115 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.901217 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.901358 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.901499 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.901623 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:09 crc kubenswrapper[4945]: I1206 08:59:09.901785 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.003837 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.003905 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.003931 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.004007 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.004032 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.004060 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-inventory\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.004095 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ceph\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.004110 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvqq6\" (UniqueName: \"kubernetes.io/projected/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-kube-api-access-nvqq6\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.004164 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.004183 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.004203 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.006386 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.015684 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.016191 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.018765 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.020012 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-inventory\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.020860 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ceph\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.036952 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.041302 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvqq6\" (UniqueName: \"kubernetes.io/projected/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-kube-api-access-nvqq6\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.041968 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.042100 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.046970 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-4pkmq\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.070765 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 08:59:10 crc kubenswrapper[4945]: I1206 08:59:10.666194 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-4pkmq"] Dec 06 08:59:10 crc kubenswrapper[4945]: W1206 08:59:10.666446 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd03da9d8_fe0b_4894_806d_2d0f15de1ea1.slice/crio-27d1a74e3c591a354ed44a7eb0a222f6b3faaeae657991301f55fed8fa9dc735 WatchSource:0}: Error finding container 27d1a74e3c591a354ed44a7eb0a222f6b3faaeae657991301f55fed8fa9dc735: Status 404 returned error can't find the container with id 27d1a74e3c591a354ed44a7eb0a222f6b3faaeae657991301f55fed8fa9dc735 Dec 06 08:59:11 crc kubenswrapper[4945]: I1206 08:59:11.638622 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" event={"ID":"d03da9d8-fe0b-4894-806d-2d0f15de1ea1","Type":"ContainerStarted","Data":"3cd09d84aaaaf5731ca3ebe87fda642961a04ecd478852b9edc44d285eac2ef5"} Dec 06 08:59:11 crc kubenswrapper[4945]: I1206 08:59:11.639174 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" event={"ID":"d03da9d8-fe0b-4894-806d-2d0f15de1ea1","Type":"ContainerStarted","Data":"27d1a74e3c591a354ed44a7eb0a222f6b3faaeae657991301f55fed8fa9dc735"} Dec 06 08:59:11 crc kubenswrapper[4945]: I1206 08:59:11.659981 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" podStartSLOduration=2.1277136580000002 podStartE2EDuration="2.65996721s" podCreationTimestamp="2025-12-06 08:59:09 +0000 UTC" firstStartedPulling="2025-12-06 08:59:10.673376538 +0000 UTC m=+7564.128237582" lastFinishedPulling="2025-12-06 08:59:11.20563009 +0000 UTC m=+7564.660491134" observedRunningTime="2025-12-06 08:59:11.655758547 +0000 UTC m=+7565.110619591" watchObservedRunningTime="2025-12-06 08:59:11.65996721 +0000 UTC m=+7565.114828254" Dec 06 08:59:13 crc kubenswrapper[4945]: I1206 08:59:13.953613 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:59:13 crc kubenswrapper[4945]: E1206 08:59:13.954155 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:59:14 crc kubenswrapper[4945]: I1206 08:59:14.867090 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6rc7q"] Dec 06 08:59:14 crc kubenswrapper[4945]: I1206 08:59:14.870188 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:14 crc kubenswrapper[4945]: I1206 08:59:14.903619 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6rc7q"] Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.074209 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-utilities\") pod \"certified-operators-6rc7q\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.074352 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-catalog-content\") pod \"certified-operators-6rc7q\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.074455 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d97q\" (UniqueName: \"kubernetes.io/projected/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-kube-api-access-9d97q\") pod \"certified-operators-6rc7q\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.176787 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-catalog-content\") pod \"certified-operators-6rc7q\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.177144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d97q\" (UniqueName: \"kubernetes.io/projected/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-kube-api-access-9d97q\") pod \"certified-operators-6rc7q\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.177444 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-catalog-content\") pod \"certified-operators-6rc7q\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.177592 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-utilities\") pod \"certified-operators-6rc7q\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.177998 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-utilities\") pod \"certified-operators-6rc7q\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.202100 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d97q\" (UniqueName: \"kubernetes.io/projected/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-kube-api-access-9d97q\") pod \"certified-operators-6rc7q\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.203693 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:15 crc kubenswrapper[4945]: W1206 08:59:15.757081 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45b67b56_0b39_42d8_a1aa_2d0f6a2666ea.slice/crio-82bf0f17fefa15cd4516140d0ea39e741cba61542db3ebc9720b8d83599f763a WatchSource:0}: Error finding container 82bf0f17fefa15cd4516140d0ea39e741cba61542db3ebc9720b8d83599f763a: Status 404 returned error can't find the container with id 82bf0f17fefa15cd4516140d0ea39e741cba61542db3ebc9720b8d83599f763a Dec 06 08:59:15 crc kubenswrapper[4945]: I1206 08:59:15.757838 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6rc7q"] Dec 06 08:59:16 crc kubenswrapper[4945]: I1206 08:59:16.705643 4945 generic.go:334] "Generic (PLEG): container finished" podID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerID="f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b" exitCode=0 Dec 06 08:59:16 crc kubenswrapper[4945]: I1206 08:59:16.706158 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rc7q" event={"ID":"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea","Type":"ContainerDied","Data":"f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b"} Dec 06 08:59:16 crc kubenswrapper[4945]: I1206 08:59:16.706233 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rc7q" event={"ID":"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea","Type":"ContainerStarted","Data":"82bf0f17fefa15cd4516140d0ea39e741cba61542db3ebc9720b8d83599f763a"} Dec 06 08:59:17 crc kubenswrapper[4945]: I1206 08:59:17.715843 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rc7q" event={"ID":"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea","Type":"ContainerStarted","Data":"36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1"} Dec 06 08:59:18 crc kubenswrapper[4945]: I1206 08:59:18.726526 4945 generic.go:334] "Generic (PLEG): container finished" podID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerID="36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1" exitCode=0 Dec 06 08:59:18 crc kubenswrapper[4945]: I1206 08:59:18.726622 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rc7q" event={"ID":"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea","Type":"ContainerDied","Data":"36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1"} Dec 06 08:59:19 crc kubenswrapper[4945]: I1206 08:59:19.736041 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rc7q" event={"ID":"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea","Type":"ContainerStarted","Data":"17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5"} Dec 06 08:59:19 crc kubenswrapper[4945]: I1206 08:59:19.761926 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6rc7q" podStartSLOduration=3.36292614 podStartE2EDuration="5.7619055s" podCreationTimestamp="2025-12-06 08:59:14 +0000 UTC" firstStartedPulling="2025-12-06 08:59:16.708894911 +0000 UTC m=+7570.163755955" lastFinishedPulling="2025-12-06 08:59:19.107874271 +0000 UTC m=+7572.562735315" observedRunningTime="2025-12-06 08:59:19.752104088 +0000 UTC m=+7573.206965132" watchObservedRunningTime="2025-12-06 08:59:19.7619055 +0000 UTC m=+7573.216766544" Dec 06 08:59:24 crc kubenswrapper[4945]: I1206 08:59:24.953252 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:59:24 crc kubenswrapper[4945]: E1206 08:59:24.954053 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:59:25 crc kubenswrapper[4945]: I1206 08:59:25.204883 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:25 crc kubenswrapper[4945]: I1206 08:59:25.204944 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:25 crc kubenswrapper[4945]: I1206 08:59:25.257838 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:25 crc kubenswrapper[4945]: I1206 08:59:25.841382 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:25 crc kubenswrapper[4945]: I1206 08:59:25.887374 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6rc7q"] Dec 06 08:59:27 crc kubenswrapper[4945]: I1206 08:59:27.914829 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6rc7q" podUID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerName="registry-server" containerID="cri-o://17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5" gracePeriod=2 Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.718745 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.838990 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-catalog-content\") pod \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.839121 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d97q\" (UniqueName: \"kubernetes.io/projected/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-kube-api-access-9d97q\") pod \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.839187 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-utilities\") pod \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\" (UID: \"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea\") " Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.840248 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-utilities" (OuterVolumeSpecName: "utilities") pod "45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" (UID: "45b67b56-0b39-42d8-a1aa-2d0f6a2666ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.845878 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-kube-api-access-9d97q" (OuterVolumeSpecName: "kube-api-access-9d97q") pod "45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" (UID: "45b67b56-0b39-42d8-a1aa-2d0f6a2666ea"). InnerVolumeSpecName "kube-api-access-9d97q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.893421 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" (UID: "45b67b56-0b39-42d8-a1aa-2d0f6a2666ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.926507 4945 generic.go:334] "Generic (PLEG): container finished" podID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerID="17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5" exitCode=0 Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.926560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rc7q" event={"ID":"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea","Type":"ContainerDied","Data":"17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5"} Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.926600 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rc7q" event={"ID":"45b67b56-0b39-42d8-a1aa-2d0f6a2666ea","Type":"ContainerDied","Data":"82bf0f17fefa15cd4516140d0ea39e741cba61542db3ebc9720b8d83599f763a"} Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.926613 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6rc7q" Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.926620 4945 scope.go:117] "RemoveContainer" containerID="17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5" Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.942467 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.942492 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d97q\" (UniqueName: \"kubernetes.io/projected/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-kube-api-access-9d97q\") on node \"crc\" DevicePath \"\"" Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.942502 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.958200 4945 scope.go:117] "RemoveContainer" containerID="36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1" Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.975914 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6rc7q"] Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.985349 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6rc7q"] Dec 06 08:59:28 crc kubenswrapper[4945]: I1206 08:59:28.997404 4945 scope.go:117] "RemoveContainer" containerID="f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b" Dec 06 08:59:29 crc kubenswrapper[4945]: I1206 08:59:29.031796 4945 scope.go:117] "RemoveContainer" containerID="17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5" Dec 06 08:59:29 crc kubenswrapper[4945]: E1206 08:59:29.032803 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5\": container with ID starting with 17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5 not found: ID does not exist" containerID="17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5" Dec 06 08:59:29 crc kubenswrapper[4945]: I1206 08:59:29.032859 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5"} err="failed to get container status \"17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5\": rpc error: code = NotFound desc = could not find container \"17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5\": container with ID starting with 17096f548a3961c7501a4b8557d3ee245e965ef7296bc01de3ec7e3d4b8660f5 not found: ID does not exist" Dec 06 08:59:29 crc kubenswrapper[4945]: I1206 08:59:29.032893 4945 scope.go:117] "RemoveContainer" containerID="36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1" Dec 06 08:59:29 crc kubenswrapper[4945]: E1206 08:59:29.033622 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1\": container with ID starting with 36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1 not found: ID does not exist" containerID="36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1" Dec 06 08:59:29 crc kubenswrapper[4945]: I1206 08:59:29.033679 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1"} err="failed to get container status \"36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1\": rpc error: code = NotFound desc = could not find container \"36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1\": container with ID starting with 36a0352b64288884e92c9b88c2b71ac73979690647597fc7411e255b46a1f2e1 not found: ID does not exist" Dec 06 08:59:29 crc kubenswrapper[4945]: I1206 08:59:29.033725 4945 scope.go:117] "RemoveContainer" containerID="f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b" Dec 06 08:59:29 crc kubenswrapper[4945]: E1206 08:59:29.034086 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b\": container with ID starting with f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b not found: ID does not exist" containerID="f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b" Dec 06 08:59:29 crc kubenswrapper[4945]: I1206 08:59:29.034123 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b"} err="failed to get container status \"f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b\": rpc error: code = NotFound desc = could not find container \"f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b\": container with ID starting with f6f76c7e1eb026b3f97acaf0d904702c00ed02dd440759151dec041b26b8bb1b not found: ID does not exist" Dec 06 08:59:30 crc kubenswrapper[4945]: I1206 08:59:30.977112 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" path="/var/lib/kubelet/pods/45b67b56-0b39-42d8-a1aa-2d0f6a2666ea/volumes" Dec 06 08:59:37 crc kubenswrapper[4945]: I1206 08:59:37.953650 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:59:37 crc kubenswrapper[4945]: E1206 08:59:37.954610 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:59:48 crc kubenswrapper[4945]: I1206 08:59:48.953762 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:59:48 crc kubenswrapper[4945]: E1206 08:59:48.954593 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 08:59:59 crc kubenswrapper[4945]: I1206 08:59:59.953273 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 08:59:59 crc kubenswrapper[4945]: E1206 08:59:59.955503 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.163642 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj"] Dec 06 09:00:00 crc kubenswrapper[4945]: E1206 09:00:00.164192 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerName="extract-content" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.164218 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerName="extract-content" Dec 06 09:00:00 crc kubenswrapper[4945]: E1206 09:00:00.164242 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerName="registry-server" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.164251 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerName="registry-server" Dec 06 09:00:00 crc kubenswrapper[4945]: E1206 09:00:00.164315 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerName="extract-utilities" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.164326 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerName="extract-utilities" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.164565 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="45b67b56-0b39-42d8-a1aa-2d0f6a2666ea" containerName="registry-server" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.165629 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.168397 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.168591 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.174357 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj"] Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.356045 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46722c9b-93e7-4d89-9de7-190570bd94a5-secret-volume\") pod \"collect-profiles-29416860-t29bj\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.356440 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46722c9b-93e7-4d89-9de7-190570bd94a5-config-volume\") pod \"collect-profiles-29416860-t29bj\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.356485 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbvcf\" (UniqueName: \"kubernetes.io/projected/46722c9b-93e7-4d89-9de7-190570bd94a5-kube-api-access-jbvcf\") pod \"collect-profiles-29416860-t29bj\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.457884 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbvcf\" (UniqueName: \"kubernetes.io/projected/46722c9b-93e7-4d89-9de7-190570bd94a5-kube-api-access-jbvcf\") pod \"collect-profiles-29416860-t29bj\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.458070 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46722c9b-93e7-4d89-9de7-190570bd94a5-secret-volume\") pod \"collect-profiles-29416860-t29bj\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.458112 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46722c9b-93e7-4d89-9de7-190570bd94a5-config-volume\") pod \"collect-profiles-29416860-t29bj\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.459030 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46722c9b-93e7-4d89-9de7-190570bd94a5-config-volume\") pod \"collect-profiles-29416860-t29bj\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.464044 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46722c9b-93e7-4d89-9de7-190570bd94a5-secret-volume\") pod \"collect-profiles-29416860-t29bj\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.475620 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbvcf\" (UniqueName: \"kubernetes.io/projected/46722c9b-93e7-4d89-9de7-190570bd94a5-kube-api-access-jbvcf\") pod \"collect-profiles-29416860-t29bj\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:00 crc kubenswrapper[4945]: I1206 09:00:00.499947 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:01 crc kubenswrapper[4945]: I1206 09:00:01.001556 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj"] Dec 06 09:00:01 crc kubenswrapper[4945]: I1206 09:00:01.271862 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" event={"ID":"46722c9b-93e7-4d89-9de7-190570bd94a5","Type":"ContainerStarted","Data":"32b8a3775cdb1a9722c999fa6ec938af3fbf2b6d1c5be738e83375f9a21d37c7"} Dec 06 09:00:01 crc kubenswrapper[4945]: I1206 09:00:01.272199 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" event={"ID":"46722c9b-93e7-4d89-9de7-190570bd94a5","Type":"ContainerStarted","Data":"893a66120f0cf07f4185c7e17b40f18d38bcadbec120139d82c63aaf3ccd3459"} Dec 06 09:00:01 crc kubenswrapper[4945]: I1206 09:00:01.283510 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" podStartSLOduration=1.2834904 podStartE2EDuration="1.2834904s" podCreationTimestamp="2025-12-06 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 09:00:01.282516984 +0000 UTC m=+7614.737378028" watchObservedRunningTime="2025-12-06 09:00:01.2834904 +0000 UTC m=+7614.738351444" Dec 06 09:00:02 crc kubenswrapper[4945]: I1206 09:00:02.293424 4945 generic.go:334] "Generic (PLEG): container finished" podID="46722c9b-93e7-4d89-9de7-190570bd94a5" containerID="32b8a3775cdb1a9722c999fa6ec938af3fbf2b6d1c5be738e83375f9a21d37c7" exitCode=0 Dec 06 09:00:02 crc kubenswrapper[4945]: I1206 09:00:02.293493 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" event={"ID":"46722c9b-93e7-4d89-9de7-190570bd94a5","Type":"ContainerDied","Data":"32b8a3775cdb1a9722c999fa6ec938af3fbf2b6d1c5be738e83375f9a21d37c7"} Dec 06 09:00:03 crc kubenswrapper[4945]: I1206 09:00:03.742413 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:03 crc kubenswrapper[4945]: I1206 09:00:03.829353 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46722c9b-93e7-4d89-9de7-190570bd94a5-secret-volume\") pod \"46722c9b-93e7-4d89-9de7-190570bd94a5\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " Dec 06 09:00:03 crc kubenswrapper[4945]: I1206 09:00:03.829471 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46722c9b-93e7-4d89-9de7-190570bd94a5-config-volume\") pod \"46722c9b-93e7-4d89-9de7-190570bd94a5\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " Dec 06 09:00:03 crc kubenswrapper[4945]: I1206 09:00:03.829556 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbvcf\" (UniqueName: \"kubernetes.io/projected/46722c9b-93e7-4d89-9de7-190570bd94a5-kube-api-access-jbvcf\") pod \"46722c9b-93e7-4d89-9de7-190570bd94a5\" (UID: \"46722c9b-93e7-4d89-9de7-190570bd94a5\") " Dec 06 09:00:03 crc kubenswrapper[4945]: I1206 09:00:03.830658 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46722c9b-93e7-4d89-9de7-190570bd94a5-config-volume" (OuterVolumeSpecName: "config-volume") pod "46722c9b-93e7-4d89-9de7-190570bd94a5" (UID: "46722c9b-93e7-4d89-9de7-190570bd94a5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 09:00:03 crc kubenswrapper[4945]: I1206 09:00:03.835047 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46722c9b-93e7-4d89-9de7-190570bd94a5-kube-api-access-jbvcf" (OuterVolumeSpecName: "kube-api-access-jbvcf") pod "46722c9b-93e7-4d89-9de7-190570bd94a5" (UID: "46722c9b-93e7-4d89-9de7-190570bd94a5"). InnerVolumeSpecName "kube-api-access-jbvcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:00:03 crc kubenswrapper[4945]: I1206 09:00:03.836213 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46722c9b-93e7-4d89-9de7-190570bd94a5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "46722c9b-93e7-4d89-9de7-190570bd94a5" (UID: "46722c9b-93e7-4d89-9de7-190570bd94a5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:00:03 crc kubenswrapper[4945]: I1206 09:00:03.931496 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46722c9b-93e7-4d89-9de7-190570bd94a5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 09:00:03 crc kubenswrapper[4945]: I1206 09:00:03.931528 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46722c9b-93e7-4d89-9de7-190570bd94a5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 09:00:03 crc kubenswrapper[4945]: I1206 09:00:03.931538 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbvcf\" (UniqueName: \"kubernetes.io/projected/46722c9b-93e7-4d89-9de7-190570bd94a5-kube-api-access-jbvcf\") on node \"crc\" DevicePath \"\"" Dec 06 09:00:04 crc kubenswrapper[4945]: I1206 09:00:04.331157 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" event={"ID":"46722c9b-93e7-4d89-9de7-190570bd94a5","Type":"ContainerDied","Data":"893a66120f0cf07f4185c7e17b40f18d38bcadbec120139d82c63aaf3ccd3459"} Dec 06 09:00:04 crc kubenswrapper[4945]: I1206 09:00:04.331595 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="893a66120f0cf07f4185c7e17b40f18d38bcadbec120139d82c63aaf3ccd3459" Dec 06 09:00:04 crc kubenswrapper[4945]: I1206 09:00:04.331198 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj" Dec 06 09:00:04 crc kubenswrapper[4945]: I1206 09:00:04.369438 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j"] Dec 06 09:00:04 crc kubenswrapper[4945]: I1206 09:00:04.378773 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416815-8x25j"] Dec 06 09:00:04 crc kubenswrapper[4945]: I1206 09:00:04.970207 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e22d9737-bf91-4dd2-8262-6e43b23ff1d2" path="/var/lib/kubelet/pods/e22d9737-bf91-4dd2-8262-6e43b23ff1d2/volumes" Dec 06 09:00:14 crc kubenswrapper[4945]: I1206 09:00:14.954142 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:00:14 crc kubenswrapper[4945]: E1206 09:00:14.955195 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:00:26 crc kubenswrapper[4945]: I1206 09:00:26.960338 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:00:26 crc kubenswrapper[4945]: E1206 09:00:26.962482 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:00:29 crc kubenswrapper[4945]: I1206 09:00:29.166820 4945 scope.go:117] "RemoveContainer" containerID="d310a26a2941cadf470ac4112accce96bf8c8df55143423cac55099fc17c0107" Dec 06 09:00:37 crc kubenswrapper[4945]: I1206 09:00:37.953510 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:00:37 crc kubenswrapper[4945]: E1206 09:00:37.954333 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:00:52 crc kubenswrapper[4945]: I1206 09:00:52.953862 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:00:52 crc kubenswrapper[4945]: E1206 09:00:52.954718 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.160154 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29416861-bj6jl"] Dec 06 09:01:00 crc kubenswrapper[4945]: E1206 09:01:00.161305 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46722c9b-93e7-4d89-9de7-190570bd94a5" containerName="collect-profiles" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.161322 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="46722c9b-93e7-4d89-9de7-190570bd94a5" containerName="collect-profiles" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.161598 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="46722c9b-93e7-4d89-9de7-190570bd94a5" containerName="collect-profiles" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.162494 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.185921 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29416861-bj6jl"] Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.315989 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnqn6\" (UniqueName: \"kubernetes.io/projected/f5127279-51b5-4c93-9ca9-f1d409c09aba-kube-api-access-jnqn6\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.316130 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-fernet-keys\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.316192 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-combined-ca-bundle\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.316235 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-config-data\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.417551 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-fernet-keys\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.417674 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-combined-ca-bundle\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.417707 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-config-data\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.417805 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnqn6\" (UniqueName: \"kubernetes.io/projected/f5127279-51b5-4c93-9ca9-f1d409c09aba-kube-api-access-jnqn6\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.424705 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-combined-ca-bundle\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.425027 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-config-data\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.425470 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-fernet-keys\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.434217 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnqn6\" (UniqueName: \"kubernetes.io/projected/f5127279-51b5-4c93-9ca9-f1d409c09aba-kube-api-access-jnqn6\") pod \"keystone-cron-29416861-bj6jl\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.484522 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:00 crc kubenswrapper[4945]: I1206 09:01:00.945594 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29416861-bj6jl"] Dec 06 09:01:01 crc kubenswrapper[4945]: I1206 09:01:01.863477 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416861-bj6jl" event={"ID":"f5127279-51b5-4c93-9ca9-f1d409c09aba","Type":"ContainerStarted","Data":"c7a69e8c07256ebc0b5edf886ac2421161b7d5345ca8a279d39b0433f8617477"} Dec 06 09:01:01 crc kubenswrapper[4945]: I1206 09:01:01.863743 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416861-bj6jl" event={"ID":"f5127279-51b5-4c93-9ca9-f1d409c09aba","Type":"ContainerStarted","Data":"c15894dae5be2dc294700df02c271dc1823537cdc556851db31c97c670ae86d8"} Dec 06 09:01:01 crc kubenswrapper[4945]: I1206 09:01:01.878702 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29416861-bj6jl" podStartSLOduration=1.878679939 podStartE2EDuration="1.878679939s" podCreationTimestamp="2025-12-06 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 09:01:01.8768332 +0000 UTC m=+7675.331694254" watchObservedRunningTime="2025-12-06 09:01:01.878679939 +0000 UTC m=+7675.333540983" Dec 06 09:01:03 crc kubenswrapper[4945]: I1206 09:01:03.883365 4945 generic.go:334] "Generic (PLEG): container finished" podID="f5127279-51b5-4c93-9ca9-f1d409c09aba" containerID="c7a69e8c07256ebc0b5edf886ac2421161b7d5345ca8a279d39b0433f8617477" exitCode=0 Dec 06 09:01:03 crc kubenswrapper[4945]: I1206 09:01:03.883706 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416861-bj6jl" event={"ID":"f5127279-51b5-4c93-9ca9-f1d409c09aba","Type":"ContainerDied","Data":"c7a69e8c07256ebc0b5edf886ac2421161b7d5345ca8a279d39b0433f8617477"} Dec 06 09:01:03 crc kubenswrapper[4945]: I1206 09:01:03.953795 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:01:03 crc kubenswrapper[4945]: E1206 09:01:03.954083 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.299944 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.421159 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-fernet-keys\") pod \"f5127279-51b5-4c93-9ca9-f1d409c09aba\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.421340 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnqn6\" (UniqueName: \"kubernetes.io/projected/f5127279-51b5-4c93-9ca9-f1d409c09aba-kube-api-access-jnqn6\") pod \"f5127279-51b5-4c93-9ca9-f1d409c09aba\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.421477 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-config-data\") pod \"f5127279-51b5-4c93-9ca9-f1d409c09aba\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.421515 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-combined-ca-bundle\") pod \"f5127279-51b5-4c93-9ca9-f1d409c09aba\" (UID: \"f5127279-51b5-4c93-9ca9-f1d409c09aba\") " Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.427156 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5127279-51b5-4c93-9ca9-f1d409c09aba-kube-api-access-jnqn6" (OuterVolumeSpecName: "kube-api-access-jnqn6") pod "f5127279-51b5-4c93-9ca9-f1d409c09aba" (UID: "f5127279-51b5-4c93-9ca9-f1d409c09aba"). InnerVolumeSpecName "kube-api-access-jnqn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.427179 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f5127279-51b5-4c93-9ca9-f1d409c09aba" (UID: "f5127279-51b5-4c93-9ca9-f1d409c09aba"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.450839 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5127279-51b5-4c93-9ca9-f1d409c09aba" (UID: "f5127279-51b5-4c93-9ca9-f1d409c09aba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.473353 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-config-data" (OuterVolumeSpecName: "config-data") pod "f5127279-51b5-4c93-9ca9-f1d409c09aba" (UID: "f5127279-51b5-4c93-9ca9-f1d409c09aba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.524134 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.524182 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnqn6\" (UniqueName: \"kubernetes.io/projected/f5127279-51b5-4c93-9ca9-f1d409c09aba-kube-api-access-jnqn6\") on node \"crc\" DevicePath \"\"" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.524194 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.524202 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5127279-51b5-4c93-9ca9-f1d409c09aba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.906612 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416861-bj6jl" event={"ID":"f5127279-51b5-4c93-9ca9-f1d409c09aba","Type":"ContainerDied","Data":"c15894dae5be2dc294700df02c271dc1823537cdc556851db31c97c670ae86d8"} Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.906982 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c15894dae5be2dc294700df02c271dc1823537cdc556851db31c97c670ae86d8" Dec 06 09:01:05 crc kubenswrapper[4945]: I1206 09:01:05.906694 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416861-bj6jl" Dec 06 09:01:17 crc kubenswrapper[4945]: I1206 09:01:17.954171 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:01:17 crc kubenswrapper[4945]: E1206 09:01:17.955092 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:01:30 crc kubenswrapper[4945]: I1206 09:01:30.953631 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:01:30 crc kubenswrapper[4945]: E1206 09:01:30.954529 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:01:45 crc kubenswrapper[4945]: I1206 09:01:45.954164 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:01:45 crc kubenswrapper[4945]: E1206 09:01:45.957951 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:01:56 crc kubenswrapper[4945]: I1206 09:01:56.960148 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:01:56 crc kubenswrapper[4945]: E1206 09:01:56.961105 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:02:07 crc kubenswrapper[4945]: I1206 09:02:07.954321 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:02:07 crc kubenswrapper[4945]: E1206 09:02:07.955077 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:02:19 crc kubenswrapper[4945]: I1206 09:02:19.662084 4945 generic.go:334] "Generic (PLEG): container finished" podID="d03da9d8-fe0b-4894-806d-2d0f15de1ea1" containerID="3cd09d84aaaaf5731ca3ebe87fda642961a04ecd478852b9edc44d285eac2ef5" exitCode=0 Dec 06 09:02:19 crc kubenswrapper[4945]: I1206 09:02:19.662685 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" event={"ID":"d03da9d8-fe0b-4894-806d-2d0f15de1ea1","Type":"ContainerDied","Data":"3cd09d84aaaaf5731ca3ebe87fda642961a04ecd478852b9edc44d285eac2ef5"} Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.100786 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.205925 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ssh-key\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.206268 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ceph\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.206530 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-0\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.206670 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvqq6\" (UniqueName: \"kubernetes.io/projected/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-kube-api-access-nvqq6\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.206769 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-1\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.206911 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-0\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.207025 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-inventory\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.207131 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-1\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.207292 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-combined-ca-bundle\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.207396 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-1\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.207537 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-0\") pod \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\" (UID: \"d03da9d8-fe0b-4894-806d-2d0f15de1ea1\") " Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.225359 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ceph" (OuterVolumeSpecName: "ceph") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.226848 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-kube-api-access-nvqq6" (OuterVolumeSpecName: "kube-api-access-nvqq6") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "kube-api-access-nvqq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.228027 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.236016 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.237039 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.241001 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-inventory" (OuterVolumeSpecName: "inventory") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.243828 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.248109 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.249580 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.251123 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.271795 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "d03da9d8-fe0b-4894-806d-2d0f15de1ea1" (UID: "d03da9d8-fe0b-4894-806d-2d0f15de1ea1"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.309680 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvqq6\" (UniqueName: \"kubernetes.io/projected/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-kube-api-access-nvqq6\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.309996 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.310006 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.310016 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.310026 4945 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.310034 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.310045 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.310053 4945 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.310061 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.310068 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.310078 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d03da9d8-fe0b-4894-806d-2d0f15de1ea1-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.685398 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" event={"ID":"d03da9d8-fe0b-4894-806d-2d0f15de1ea1","Type":"ContainerDied","Data":"27d1a74e3c591a354ed44a7eb0a222f6b3faaeae657991301f55fed8fa9dc735"} Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.685447 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27d1a74e3c591a354ed44a7eb0a222f6b3faaeae657991301f55fed8fa9dc735" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.685535 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-4pkmq" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.806241 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-rlgng"] Dec 06 09:02:21 crc kubenswrapper[4945]: E1206 09:02:21.806735 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5127279-51b5-4c93-9ca9-f1d409c09aba" containerName="keystone-cron" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.806757 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5127279-51b5-4c93-9ca9-f1d409c09aba" containerName="keystone-cron" Dec 06 09:02:21 crc kubenswrapper[4945]: E1206 09:02:21.806800 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d03da9d8-fe0b-4894-806d-2d0f15de1ea1" containerName="nova-cell1-openstack-openstack-cell1" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.806808 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d03da9d8-fe0b-4894-806d-2d0f15de1ea1" containerName="nova-cell1-openstack-openstack-cell1" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.806987 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5127279-51b5-4c93-9ca9-f1d409c09aba" containerName="keystone-cron" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.807017 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d03da9d8-fe0b-4894-806d-2d0f15de1ea1" containerName="nova-cell1-openstack-openstack-cell1" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.807794 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.810024 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.810304 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.810410 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.810414 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.810472 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.817611 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-rlgng"] Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.920976 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ssh-key\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.921027 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbnhl\" (UniqueName: \"kubernetes.io/projected/373f7931-312f-4269-932e-53fc9a2e1e6d-kube-api-access-sbnhl\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.921098 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.921396 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-inventory\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.921496 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.921706 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.921818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.921937 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceph\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:21 crc kubenswrapper[4945]: I1206 09:02:21.953687 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.024005 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ssh-key\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.024051 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbnhl\" (UniqueName: \"kubernetes.io/projected/373f7931-312f-4269-932e-53fc9a2e1e6d-kube-api-access-sbnhl\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.024114 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.024186 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-inventory\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.024219 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.024326 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.024366 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.024428 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceph\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.029918 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.029918 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-inventory\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.030054 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ssh-key\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.030146 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.030523 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.031040 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceph\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.031498 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.042191 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbnhl\" (UniqueName: \"kubernetes.io/projected/373f7931-312f-4269-932e-53fc9a2e1e6d-kube-api-access-sbnhl\") pod \"telemetry-openstack-openstack-cell1-rlgng\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.131124 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.680472 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-rlgng"] Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.706428 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-rlgng" event={"ID":"373f7931-312f-4269-932e-53fc9a2e1e6d","Type":"ContainerStarted","Data":"8d4533007b5b8ad5e0415775dcfc4560c1d660bc3475f789b3f64280d520cfdd"} Dec 06 09:02:22 crc kubenswrapper[4945]: I1206 09:02:22.709372 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"a6e58a6773016841d512f08b6f03902cd6801ad8709782693c3094dafd869b22"} Dec 06 09:02:23 crc kubenswrapper[4945]: I1206 09:02:23.730693 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-rlgng" event={"ID":"373f7931-312f-4269-932e-53fc9a2e1e6d","Type":"ContainerStarted","Data":"a5d48a33f97642a2dcaf857b07790208dd6a4e1fb6da3a9adae934be7fa701f1"} Dec 06 09:02:23 crc kubenswrapper[4945]: I1206 09:02:23.767545 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-rlgng" podStartSLOduration=2.246097797 podStartE2EDuration="2.767350556s" podCreationTimestamp="2025-12-06 09:02:21 +0000 UTC" firstStartedPulling="2025-12-06 09:02:22.680775521 +0000 UTC m=+7756.135636555" lastFinishedPulling="2025-12-06 09:02:23.20202827 +0000 UTC m=+7756.656889314" observedRunningTime="2025-12-06 09:02:23.751256545 +0000 UTC m=+7757.206117589" watchObservedRunningTime="2025-12-06 09:02:23.767350556 +0000 UTC m=+7757.222211600" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.174045 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9mzqd"] Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.177063 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.195808 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mzqd"] Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.272797 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-utilities\") pod \"redhat-marketplace-9mzqd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.272864 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5k5x\" (UniqueName: \"kubernetes.io/projected/3133668c-8eb6-474a-9865-3e44910ab5fd-kube-api-access-f5k5x\") pod \"redhat-marketplace-9mzqd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.272907 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-catalog-content\") pod \"redhat-marketplace-9mzqd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.375114 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-utilities\") pod \"redhat-marketplace-9mzqd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.375438 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5k5x\" (UniqueName: \"kubernetes.io/projected/3133668c-8eb6-474a-9865-3e44910ab5fd-kube-api-access-f5k5x\") pod \"redhat-marketplace-9mzqd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.375471 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-catalog-content\") pod \"redhat-marketplace-9mzqd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.375641 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-utilities\") pod \"redhat-marketplace-9mzqd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.375827 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-catalog-content\") pod \"redhat-marketplace-9mzqd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.403035 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5k5x\" (UniqueName: \"kubernetes.io/projected/3133668c-8eb6-474a-9865-3e44910ab5fd-kube-api-access-f5k5x\") pod \"redhat-marketplace-9mzqd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:24 crc kubenswrapper[4945]: I1206 09:02:24.502305 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:25 crc kubenswrapper[4945]: I1206 09:02:25.216976 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mzqd"] Dec 06 09:02:25 crc kubenswrapper[4945]: I1206 09:02:25.748300 4945 generic.go:334] "Generic (PLEG): container finished" podID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerID="a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369" exitCode=0 Dec 06 09:02:25 crc kubenswrapper[4945]: I1206 09:02:25.748363 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mzqd" event={"ID":"3133668c-8eb6-474a-9865-3e44910ab5fd","Type":"ContainerDied","Data":"a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369"} Dec 06 09:02:25 crc kubenswrapper[4945]: I1206 09:02:25.749200 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mzqd" event={"ID":"3133668c-8eb6-474a-9865-3e44910ab5fd","Type":"ContainerStarted","Data":"765042686daae498505b9ddccc585674e0e4a87187222592c9f649a43bcad2dd"} Dec 06 09:02:25 crc kubenswrapper[4945]: I1206 09:02:25.750081 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 09:02:27 crc kubenswrapper[4945]: I1206 09:02:27.766845 4945 generic.go:334] "Generic (PLEG): container finished" podID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerID="d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca" exitCode=0 Dec 06 09:02:27 crc kubenswrapper[4945]: I1206 09:02:27.766903 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mzqd" event={"ID":"3133668c-8eb6-474a-9865-3e44910ab5fd","Type":"ContainerDied","Data":"d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca"} Dec 06 09:02:28 crc kubenswrapper[4945]: I1206 09:02:28.781984 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mzqd" event={"ID":"3133668c-8eb6-474a-9865-3e44910ab5fd","Type":"ContainerStarted","Data":"829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a"} Dec 06 09:02:28 crc kubenswrapper[4945]: I1206 09:02:28.814210 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9mzqd" podStartSLOduration=2.393044209 podStartE2EDuration="4.814184781s" podCreationTimestamp="2025-12-06 09:02:24 +0000 UTC" firstStartedPulling="2025-12-06 09:02:25.749894441 +0000 UTC m=+7759.204755485" lastFinishedPulling="2025-12-06 09:02:28.171035013 +0000 UTC m=+7761.625896057" observedRunningTime="2025-12-06 09:02:28.806114235 +0000 UTC m=+7762.260975339" watchObservedRunningTime="2025-12-06 09:02:28.814184781 +0000 UTC m=+7762.269045825" Dec 06 09:02:34 crc kubenswrapper[4945]: I1206 09:02:34.503086 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:34 crc kubenswrapper[4945]: I1206 09:02:34.503723 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:34 crc kubenswrapper[4945]: I1206 09:02:34.563266 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:34 crc kubenswrapper[4945]: I1206 09:02:34.891767 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:34 crc kubenswrapper[4945]: I1206 09:02:34.944969 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mzqd"] Dec 06 09:02:36 crc kubenswrapper[4945]: I1206 09:02:36.852181 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9mzqd" podUID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerName="registry-server" containerID="cri-o://829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a" gracePeriod=2 Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.378391 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.566027 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-catalog-content\") pod \"3133668c-8eb6-474a-9865-3e44910ab5fd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.566599 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-utilities\") pod \"3133668c-8eb6-474a-9865-3e44910ab5fd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.566681 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5k5x\" (UniqueName: \"kubernetes.io/projected/3133668c-8eb6-474a-9865-3e44910ab5fd-kube-api-access-f5k5x\") pod \"3133668c-8eb6-474a-9865-3e44910ab5fd\" (UID: \"3133668c-8eb6-474a-9865-3e44910ab5fd\") " Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.567141 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-utilities" (OuterVolumeSpecName: "utilities") pod "3133668c-8eb6-474a-9865-3e44910ab5fd" (UID: "3133668c-8eb6-474a-9865-3e44910ab5fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.567633 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.573463 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3133668c-8eb6-474a-9865-3e44910ab5fd-kube-api-access-f5k5x" (OuterVolumeSpecName: "kube-api-access-f5k5x") pod "3133668c-8eb6-474a-9865-3e44910ab5fd" (UID: "3133668c-8eb6-474a-9865-3e44910ab5fd"). InnerVolumeSpecName "kube-api-access-f5k5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.584799 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3133668c-8eb6-474a-9865-3e44910ab5fd" (UID: "3133668c-8eb6-474a-9865-3e44910ab5fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.669673 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3133668c-8eb6-474a-9865-3e44910ab5fd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.669709 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5k5x\" (UniqueName: \"kubernetes.io/projected/3133668c-8eb6-474a-9865-3e44910ab5fd-kube-api-access-f5k5x\") on node \"crc\" DevicePath \"\"" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.867453 4945 generic.go:334] "Generic (PLEG): container finished" podID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerID="829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a" exitCode=0 Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.867511 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mzqd" event={"ID":"3133668c-8eb6-474a-9865-3e44910ab5fd","Type":"ContainerDied","Data":"829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a"} Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.867547 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mzqd" event={"ID":"3133668c-8eb6-474a-9865-3e44910ab5fd","Type":"ContainerDied","Data":"765042686daae498505b9ddccc585674e0e4a87187222592c9f649a43bcad2dd"} Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.867573 4945 scope.go:117] "RemoveContainer" containerID="829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.867828 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mzqd" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.898500 4945 scope.go:117] "RemoveContainer" containerID="d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.926049 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mzqd"] Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.934496 4945 scope.go:117] "RemoveContainer" containerID="a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.938850 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mzqd"] Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.972113 4945 scope.go:117] "RemoveContainer" containerID="829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a" Dec 06 09:02:37 crc kubenswrapper[4945]: E1206 09:02:37.972582 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a\": container with ID starting with 829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a not found: ID does not exist" containerID="829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.972630 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a"} err="failed to get container status \"829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a\": rpc error: code = NotFound desc = could not find container \"829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a\": container with ID starting with 829eac73878525ba91801d6e94461277ee7dc7afe26c33229f5d134e899be91a not found: ID does not exist" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.972657 4945 scope.go:117] "RemoveContainer" containerID="d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca" Dec 06 09:02:37 crc kubenswrapper[4945]: E1206 09:02:37.972954 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca\": container with ID starting with d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca not found: ID does not exist" containerID="d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.972990 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca"} err="failed to get container status \"d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca\": rpc error: code = NotFound desc = could not find container \"d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca\": container with ID starting with d1fd2c6cea5e79dde08c739dcdad41ff864ef926a2ade6bcbd8bb5c9a13b49ca not found: ID does not exist" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.973009 4945 scope.go:117] "RemoveContainer" containerID="a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369" Dec 06 09:02:37 crc kubenswrapper[4945]: E1206 09:02:37.973267 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369\": container with ID starting with a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369 not found: ID does not exist" containerID="a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369" Dec 06 09:02:37 crc kubenswrapper[4945]: I1206 09:02:37.973301 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369"} err="failed to get container status \"a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369\": rpc error: code = NotFound desc = could not find container \"a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369\": container with ID starting with a4d8dd9278d3993f8a4caf9c33e7e2dae0a2f420e8df428955443f9d4e85c369 not found: ID does not exist" Dec 06 09:02:38 crc kubenswrapper[4945]: I1206 09:02:38.973742 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3133668c-8eb6-474a-9865-3e44910ab5fd" path="/var/lib/kubelet/pods/3133668c-8eb6-474a-9865-3e44910ab5fd/volumes" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.569954 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8xwcj"] Dec 06 09:04:25 crc kubenswrapper[4945]: E1206 09:04:25.571139 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerName="extract-content" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.571161 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerName="extract-content" Dec 06 09:04:25 crc kubenswrapper[4945]: E1206 09:04:25.571215 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerName="registry-server" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.571223 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerName="registry-server" Dec 06 09:04:25 crc kubenswrapper[4945]: E1206 09:04:25.571259 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerName="extract-utilities" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.571270 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerName="extract-utilities" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.571550 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3133668c-8eb6-474a-9865-3e44910ab5fd" containerName="registry-server" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.573455 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.606194 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8xwcj"] Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.688777 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59phh\" (UniqueName: \"kubernetes.io/projected/085bfca7-3891-4144-8456-4d394d2b0bcc-kube-api-access-59phh\") pod \"community-operators-8xwcj\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.688833 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-utilities\") pod \"community-operators-8xwcj\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.688874 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-catalog-content\") pod \"community-operators-8xwcj\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.790265 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59phh\" (UniqueName: \"kubernetes.io/projected/085bfca7-3891-4144-8456-4d394d2b0bcc-kube-api-access-59phh\") pod \"community-operators-8xwcj\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.790334 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-utilities\") pod \"community-operators-8xwcj\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.790376 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-catalog-content\") pod \"community-operators-8xwcj\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.790878 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-utilities\") pod \"community-operators-8xwcj\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.790948 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-catalog-content\") pod \"community-operators-8xwcj\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.810449 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59phh\" (UniqueName: \"kubernetes.io/projected/085bfca7-3891-4144-8456-4d394d2b0bcc-kube-api-access-59phh\") pod \"community-operators-8xwcj\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:25 crc kubenswrapper[4945]: I1206 09:04:25.899836 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:26 crc kubenswrapper[4945]: I1206 09:04:26.385063 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8xwcj"] Dec 06 09:04:27 crc kubenswrapper[4945]: I1206 09:04:27.052180 4945 generic.go:334] "Generic (PLEG): container finished" podID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerID="14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed" exitCode=0 Dec 06 09:04:27 crc kubenswrapper[4945]: I1206 09:04:27.052550 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xwcj" event={"ID":"085bfca7-3891-4144-8456-4d394d2b0bcc","Type":"ContainerDied","Data":"14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed"} Dec 06 09:04:27 crc kubenswrapper[4945]: I1206 09:04:27.052578 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xwcj" event={"ID":"085bfca7-3891-4144-8456-4d394d2b0bcc","Type":"ContainerStarted","Data":"37d232fe2c82c9bb59a4ca81373b943e7e54e3c2fb9590e5e2da65af6ea3e61f"} Dec 06 09:04:29 crc kubenswrapper[4945]: I1206 09:04:29.070932 4945 generic.go:334] "Generic (PLEG): container finished" podID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerID="88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724" exitCode=0 Dec 06 09:04:29 crc kubenswrapper[4945]: I1206 09:04:29.071021 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xwcj" event={"ID":"085bfca7-3891-4144-8456-4d394d2b0bcc","Type":"ContainerDied","Data":"88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724"} Dec 06 09:04:30 crc kubenswrapper[4945]: I1206 09:04:30.083065 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xwcj" event={"ID":"085bfca7-3891-4144-8456-4d394d2b0bcc","Type":"ContainerStarted","Data":"6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458"} Dec 06 09:04:30 crc kubenswrapper[4945]: I1206 09:04:30.104500 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8xwcj" podStartSLOduration=2.59505927 podStartE2EDuration="5.104483073s" podCreationTimestamp="2025-12-06 09:04:25 +0000 UTC" firstStartedPulling="2025-12-06 09:04:27.054102034 +0000 UTC m=+7880.508963078" lastFinishedPulling="2025-12-06 09:04:29.563525837 +0000 UTC m=+7883.018386881" observedRunningTime="2025-12-06 09:04:30.100074665 +0000 UTC m=+7883.554935709" watchObservedRunningTime="2025-12-06 09:04:30.104483073 +0000 UTC m=+7883.559344117" Dec 06 09:04:35 crc kubenswrapper[4945]: I1206 09:04:35.900157 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:35 crc kubenswrapper[4945]: I1206 09:04:35.900750 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:35 crc kubenswrapper[4945]: I1206 09:04:35.950758 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:36 crc kubenswrapper[4945]: I1206 09:04:36.184555 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:38 crc kubenswrapper[4945]: I1206 09:04:38.795484 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:04:38 crc kubenswrapper[4945]: I1206 09:04:38.796097 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:04:39 crc kubenswrapper[4945]: I1206 09:04:39.556837 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8xwcj"] Dec 06 09:04:39 crc kubenswrapper[4945]: I1206 09:04:39.557104 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8xwcj" podUID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerName="registry-server" containerID="cri-o://6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458" gracePeriod=2 Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.071479 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.169595 4945 generic.go:334] "Generic (PLEG): container finished" podID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerID="6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458" exitCode=0 Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.169641 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xwcj" event={"ID":"085bfca7-3891-4144-8456-4d394d2b0bcc","Type":"ContainerDied","Data":"6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458"} Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.169697 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xwcj" event={"ID":"085bfca7-3891-4144-8456-4d394d2b0bcc","Type":"ContainerDied","Data":"37d232fe2c82c9bb59a4ca81373b943e7e54e3c2fb9590e5e2da65af6ea3e61f"} Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.169719 4945 scope.go:117] "RemoveContainer" containerID="6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.169720 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xwcj" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.193732 4945 scope.go:117] "RemoveContainer" containerID="88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.205757 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-utilities\") pod \"085bfca7-3891-4144-8456-4d394d2b0bcc\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.205844 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59phh\" (UniqueName: \"kubernetes.io/projected/085bfca7-3891-4144-8456-4d394d2b0bcc-kube-api-access-59phh\") pod \"085bfca7-3891-4144-8456-4d394d2b0bcc\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.205994 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-catalog-content\") pod \"085bfca7-3891-4144-8456-4d394d2b0bcc\" (UID: \"085bfca7-3891-4144-8456-4d394d2b0bcc\") " Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.206792 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-utilities" (OuterVolumeSpecName: "utilities") pod "085bfca7-3891-4144-8456-4d394d2b0bcc" (UID: "085bfca7-3891-4144-8456-4d394d2b0bcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.211366 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/085bfca7-3891-4144-8456-4d394d2b0bcc-kube-api-access-59phh" (OuterVolumeSpecName: "kube-api-access-59phh") pod "085bfca7-3891-4144-8456-4d394d2b0bcc" (UID: "085bfca7-3891-4144-8456-4d394d2b0bcc"). InnerVolumeSpecName "kube-api-access-59phh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.215484 4945 scope.go:117] "RemoveContainer" containerID="14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.270427 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "085bfca7-3891-4144-8456-4d394d2b0bcc" (UID: "085bfca7-3891-4144-8456-4d394d2b0bcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.308821 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.308866 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59phh\" (UniqueName: \"kubernetes.io/projected/085bfca7-3891-4144-8456-4d394d2b0bcc-kube-api-access-59phh\") on node \"crc\" DevicePath \"\"" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.308879 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/085bfca7-3891-4144-8456-4d394d2b0bcc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.311135 4945 scope.go:117] "RemoveContainer" containerID="6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458" Dec 06 09:04:40 crc kubenswrapper[4945]: E1206 09:04:40.311869 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458\": container with ID starting with 6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458 not found: ID does not exist" containerID="6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.311910 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458"} err="failed to get container status \"6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458\": rpc error: code = NotFound desc = could not find container \"6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458\": container with ID starting with 6e5b4ea7f48c050e2328341c211a6379c82f9809be1cb200368c4200a84c0458 not found: ID does not exist" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.311940 4945 scope.go:117] "RemoveContainer" containerID="88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724" Dec 06 09:04:40 crc kubenswrapper[4945]: E1206 09:04:40.312255 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724\": container with ID starting with 88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724 not found: ID does not exist" containerID="88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.312309 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724"} err="failed to get container status \"88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724\": rpc error: code = NotFound desc = could not find container \"88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724\": container with ID starting with 88a375d8888bcd0c8d05e792e4500464f2feb96085c7ce1ee28f1d8474d10724 not found: ID does not exist" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.312336 4945 scope.go:117] "RemoveContainer" containerID="14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed" Dec 06 09:04:40 crc kubenswrapper[4945]: E1206 09:04:40.312580 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed\": container with ID starting with 14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed not found: ID does not exist" containerID="14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.312611 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed"} err="failed to get container status \"14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed\": rpc error: code = NotFound desc = could not find container \"14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed\": container with ID starting with 14b7b72eee5f4783a7d497071fb68da0de640613c6eabac45e7e724b7be508ed not found: ID does not exist" Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.503979 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8xwcj"] Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.513687 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8xwcj"] Dec 06 09:04:40 crc kubenswrapper[4945]: I1206 09:04:40.966808 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="085bfca7-3891-4144-8456-4d394d2b0bcc" path="/var/lib/kubelet/pods/085bfca7-3891-4144-8456-4d394d2b0bcc/volumes" Dec 06 09:05:08 crc kubenswrapper[4945]: I1206 09:05:08.796098 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:05:08 crc kubenswrapper[4945]: I1206 09:05:08.796744 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:05:38 crc kubenswrapper[4945]: I1206 09:05:38.795587 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:05:38 crc kubenswrapper[4945]: I1206 09:05:38.796047 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:05:38 crc kubenswrapper[4945]: I1206 09:05:38.796092 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 09:05:38 crc kubenswrapper[4945]: I1206 09:05:38.796898 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a6e58a6773016841d512f08b6f03902cd6801ad8709782693c3094dafd869b22"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 09:05:38 crc kubenswrapper[4945]: I1206 09:05:38.796960 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://a6e58a6773016841d512f08b6f03902cd6801ad8709782693c3094dafd869b22" gracePeriod=600 Dec 06 09:05:39 crc kubenswrapper[4945]: I1206 09:05:39.349791 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="a6e58a6773016841d512f08b6f03902cd6801ad8709782693c3094dafd869b22" exitCode=0 Dec 06 09:05:39 crc kubenswrapper[4945]: I1206 09:05:39.350141 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"a6e58a6773016841d512f08b6f03902cd6801ad8709782693c3094dafd869b22"} Dec 06 09:05:39 crc kubenswrapper[4945]: I1206 09:05:39.350171 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f"} Dec 06 09:05:39 crc kubenswrapper[4945]: I1206 09:05:39.350187 4945 scope.go:117] "RemoveContainer" containerID="3613a364c719b1272917301d773b1db3cee5235ab6a4a867187bf29b67f15774" Dec 06 09:06:10 crc kubenswrapper[4945]: I1206 09:06:10.699002 4945 generic.go:334] "Generic (PLEG): container finished" podID="373f7931-312f-4269-932e-53fc9a2e1e6d" containerID="a5d48a33f97642a2dcaf857b07790208dd6a4e1fb6da3a9adae934be7fa701f1" exitCode=0 Dec 06 09:06:10 crc kubenswrapper[4945]: I1206 09:06:10.699204 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-rlgng" event={"ID":"373f7931-312f-4269-932e-53fc9a2e1e6d","Type":"ContainerDied","Data":"a5d48a33f97642a2dcaf857b07790208dd6a4e1fb6da3a9adae934be7fa701f1"} Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.136123 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.235988 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbnhl\" (UniqueName: \"kubernetes.io/projected/373f7931-312f-4269-932e-53fc9a2e1e6d-kube-api-access-sbnhl\") pod \"373f7931-312f-4269-932e-53fc9a2e1e6d\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.236050 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-telemetry-combined-ca-bundle\") pod \"373f7931-312f-4269-932e-53fc9a2e1e6d\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.236134 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-2\") pod \"373f7931-312f-4269-932e-53fc9a2e1e6d\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.236161 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ssh-key\") pod \"373f7931-312f-4269-932e-53fc9a2e1e6d\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.236176 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-0\") pod \"373f7931-312f-4269-932e-53fc9a2e1e6d\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.236244 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-inventory\") pod \"373f7931-312f-4269-932e-53fc9a2e1e6d\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.236351 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-1\") pod \"373f7931-312f-4269-932e-53fc9a2e1e6d\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.236373 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceph\") pod \"373f7931-312f-4269-932e-53fc9a2e1e6d\" (UID: \"373f7931-312f-4269-932e-53fc9a2e1e6d\") " Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.241796 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "373f7931-312f-4269-932e-53fc9a2e1e6d" (UID: "373f7931-312f-4269-932e-53fc9a2e1e6d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.241957 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceph" (OuterVolumeSpecName: "ceph") pod "373f7931-312f-4269-932e-53fc9a2e1e6d" (UID: "373f7931-312f-4269-932e-53fc9a2e1e6d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.242025 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/373f7931-312f-4269-932e-53fc9a2e1e6d-kube-api-access-sbnhl" (OuterVolumeSpecName: "kube-api-access-sbnhl") pod "373f7931-312f-4269-932e-53fc9a2e1e6d" (UID: "373f7931-312f-4269-932e-53fc9a2e1e6d"). InnerVolumeSpecName "kube-api-access-sbnhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.265413 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "373f7931-312f-4269-932e-53fc9a2e1e6d" (UID: "373f7931-312f-4269-932e-53fc9a2e1e6d"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.265627 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-inventory" (OuterVolumeSpecName: "inventory") pod "373f7931-312f-4269-932e-53fc9a2e1e6d" (UID: "373f7931-312f-4269-932e-53fc9a2e1e6d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.267415 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "373f7931-312f-4269-932e-53fc9a2e1e6d" (UID: "373f7931-312f-4269-932e-53fc9a2e1e6d"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.274381 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "373f7931-312f-4269-932e-53fc9a2e1e6d" (UID: "373f7931-312f-4269-932e-53fc9a2e1e6d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.275047 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "373f7931-312f-4269-932e-53fc9a2e1e6d" (UID: "373f7931-312f-4269-932e-53fc9a2e1e6d"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.338839 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbnhl\" (UniqueName: \"kubernetes.io/projected/373f7931-312f-4269-932e-53fc9a2e1e6d-kube-api-access-sbnhl\") on node \"crc\" DevicePath \"\"" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.338871 4945 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.338882 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.338894 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.338904 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.338915 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.338935 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.338946 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/373f7931-312f-4269-932e-53fc9a2e1e6d-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.762915 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-rlgng" event={"ID":"373f7931-312f-4269-932e-53fc9a2e1e6d","Type":"ContainerDied","Data":"8d4533007b5b8ad5e0415775dcfc4560c1d660bc3475f789b3f64280d520cfdd"} Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.763276 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d4533007b5b8ad5e0415775dcfc4560c1d660bc3475f789b3f64280d520cfdd" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.763111 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-rlgng" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.834386 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-x45jt"] Dec 06 09:06:12 crc kubenswrapper[4945]: E1206 09:06:12.834930 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373f7931-312f-4269-932e-53fc9a2e1e6d" containerName="telemetry-openstack-openstack-cell1" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.834952 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="373f7931-312f-4269-932e-53fc9a2e1e6d" containerName="telemetry-openstack-openstack-cell1" Dec 06 09:06:12 crc kubenswrapper[4945]: E1206 09:06:12.834972 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerName="registry-server" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.834982 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerName="registry-server" Dec 06 09:06:12 crc kubenswrapper[4945]: E1206 09:06:12.835011 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerName="extract-content" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.835017 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerName="extract-content" Dec 06 09:06:12 crc kubenswrapper[4945]: E1206 09:06:12.835030 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerName="extract-utilities" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.835038 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerName="extract-utilities" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.835272 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="085bfca7-3891-4144-8456-4d394d2b0bcc" containerName="registry-server" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.835316 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="373f7931-312f-4269-932e-53fc9a2e1e6d" containerName="telemetry-openstack-openstack-cell1" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.836123 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.838116 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.838665 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.839013 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.839419 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.839459 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.844674 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-x45jt"] Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.959371 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.959432 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.959465 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.959523 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.959557 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:12 crc kubenswrapper[4945]: I1206 09:06:12.959582 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qckzw\" (UniqueName: \"kubernetes.io/projected/831dbf28-be96-4229-bd44-f88bb5b34e9c-kube-api-access-qckzw\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.060980 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.061054 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.061101 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qckzw\" (UniqueName: \"kubernetes.io/projected/831dbf28-be96-4229-bd44-f88bb5b34e9c-kube-api-access-qckzw\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.061205 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.061261 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.061322 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.066102 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.066670 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.069666 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.074178 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.074624 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.081980 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qckzw\" (UniqueName: \"kubernetes.io/projected/831dbf28-be96-4229-bd44-f88bb5b34e9c-kube-api-access-qckzw\") pod \"neutron-sriov-openstack-openstack-cell1-x45jt\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.175642 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.701427 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-x45jt"] Dec 06 09:06:13 crc kubenswrapper[4945]: I1206 09:06:13.772854 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" event={"ID":"831dbf28-be96-4229-bd44-f88bb5b34e9c","Type":"ContainerStarted","Data":"b038985b18910305489223f546deecad7c960de0d1d5e7dc819d3f382dbb7d18"} Dec 06 09:06:14 crc kubenswrapper[4945]: I1206 09:06:14.782594 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" event={"ID":"831dbf28-be96-4229-bd44-f88bb5b34e9c","Type":"ContainerStarted","Data":"4e5232a8bd6a045e017b924b900ce90d64e3a14affd4aba42624d8e680d3f7c2"} Dec 06 09:06:14 crc kubenswrapper[4945]: I1206 09:06:14.805522 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" podStartSLOduration=2.189594316 podStartE2EDuration="2.805502036s" podCreationTimestamp="2025-12-06 09:06:12 +0000 UTC" firstStartedPulling="2025-12-06 09:06:13.702152882 +0000 UTC m=+7987.157013926" lastFinishedPulling="2025-12-06 09:06:14.318060602 +0000 UTC m=+7987.772921646" observedRunningTime="2025-12-06 09:06:14.79927077 +0000 UTC m=+7988.254131834" watchObservedRunningTime="2025-12-06 09:06:14.805502036 +0000 UTC m=+7988.260363080" Dec 06 09:08:08 crc kubenswrapper[4945]: I1206 09:08:08.796646 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:08:08 crc kubenswrapper[4945]: I1206 09:08:08.797400 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:08:38 crc kubenswrapper[4945]: I1206 09:08:38.795180 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:08:38 crc kubenswrapper[4945]: I1206 09:08:38.795808 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:09:08 crc kubenswrapper[4945]: I1206 09:09:08.795609 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:09:08 crc kubenswrapper[4945]: I1206 09:09:08.796141 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:09:08 crc kubenswrapper[4945]: I1206 09:09:08.796187 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 09:09:08 crc kubenswrapper[4945]: I1206 09:09:08.797111 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 09:09:08 crc kubenswrapper[4945]: I1206 09:09:08.797175 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" gracePeriod=600 Dec 06 09:09:08 crc kubenswrapper[4945]: E1206 09:09:08.919468 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:09:09 crc kubenswrapper[4945]: I1206 09:09:09.575746 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" exitCode=0 Dec 06 09:09:09 crc kubenswrapper[4945]: I1206 09:09:09.575840 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f"} Dec 06 09:09:09 crc kubenswrapper[4945]: I1206 09:09:09.576471 4945 scope.go:117] "RemoveContainer" containerID="a6e58a6773016841d512f08b6f03902cd6801ad8709782693c3094dafd869b22" Dec 06 09:09:09 crc kubenswrapper[4945]: I1206 09:09:09.577203 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:09:09 crc kubenswrapper[4945]: E1206 09:09:09.577484 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:09:22 crc kubenswrapper[4945]: I1206 09:09:22.953385 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:09:22 crc kubenswrapper[4945]: E1206 09:09:22.954201 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:09:29 crc kubenswrapper[4945]: I1206 09:09:29.806657 4945 generic.go:334] "Generic (PLEG): container finished" podID="831dbf28-be96-4229-bd44-f88bb5b34e9c" containerID="4e5232a8bd6a045e017b924b900ce90d64e3a14affd4aba42624d8e680d3f7c2" exitCode=0 Dec 06 09:09:29 crc kubenswrapper[4945]: I1206 09:09:29.806742 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" event={"ID":"831dbf28-be96-4229-bd44-f88bb5b34e9c","Type":"ContainerDied","Data":"4e5232a8bd6a045e017b924b900ce90d64e3a14affd4aba42624d8e680d3f7c2"} Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.387850 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.506493 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qckzw\" (UniqueName: \"kubernetes.io/projected/831dbf28-be96-4229-bd44-f88bb5b34e9c-kube-api-access-qckzw\") pod \"831dbf28-be96-4229-bd44-f88bb5b34e9c\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.506579 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ssh-key\") pod \"831dbf28-be96-4229-bd44-f88bb5b34e9c\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.506637 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ceph\") pod \"831dbf28-be96-4229-bd44-f88bb5b34e9c\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.506802 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-agent-neutron-config-0\") pod \"831dbf28-be96-4229-bd44-f88bb5b34e9c\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.506875 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-combined-ca-bundle\") pod \"831dbf28-be96-4229-bd44-f88bb5b34e9c\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.506921 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-inventory\") pod \"831dbf28-be96-4229-bd44-f88bb5b34e9c\" (UID: \"831dbf28-be96-4229-bd44-f88bb5b34e9c\") " Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.514390 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ceph" (OuterVolumeSpecName: "ceph") pod "831dbf28-be96-4229-bd44-f88bb5b34e9c" (UID: "831dbf28-be96-4229-bd44-f88bb5b34e9c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.515042 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "831dbf28-be96-4229-bd44-f88bb5b34e9c" (UID: "831dbf28-be96-4229-bd44-f88bb5b34e9c"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.515750 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/831dbf28-be96-4229-bd44-f88bb5b34e9c-kube-api-access-qckzw" (OuterVolumeSpecName: "kube-api-access-qckzw") pod "831dbf28-be96-4229-bd44-f88bb5b34e9c" (UID: "831dbf28-be96-4229-bd44-f88bb5b34e9c"). InnerVolumeSpecName "kube-api-access-qckzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.543813 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "831dbf28-be96-4229-bd44-f88bb5b34e9c" (UID: "831dbf28-be96-4229-bd44-f88bb5b34e9c"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.544506 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-inventory" (OuterVolumeSpecName: "inventory") pod "831dbf28-be96-4229-bd44-f88bb5b34e9c" (UID: "831dbf28-be96-4229-bd44-f88bb5b34e9c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.551739 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "831dbf28-be96-4229-bd44-f88bb5b34e9c" (UID: "831dbf28-be96-4229-bd44-f88bb5b34e9c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.610376 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qckzw\" (UniqueName: \"kubernetes.io/projected/831dbf28-be96-4229-bd44-f88bb5b34e9c-kube-api-access-qckzw\") on node \"crc\" DevicePath \"\"" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.610424 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.610439 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.610456 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.610476 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.610493 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/831dbf28-be96-4229-bd44-f88bb5b34e9c-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.835489 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" event={"ID":"831dbf28-be96-4229-bd44-f88bb5b34e9c","Type":"ContainerDied","Data":"b038985b18910305489223f546deecad7c960de0d1d5e7dc819d3f382dbb7d18"} Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.835560 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b038985b18910305489223f546deecad7c960de0d1d5e7dc819d3f382dbb7d18" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.835650 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x45jt" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.934179 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-whzbj"] Dec 06 09:09:31 crc kubenswrapper[4945]: E1206 09:09:31.935126 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="831dbf28-be96-4229-bd44-f88bb5b34e9c" containerName="neutron-sriov-openstack-openstack-cell1" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.935168 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="831dbf28-be96-4229-bd44-f88bb5b34e9c" containerName="neutron-sriov-openstack-openstack-cell1" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.935492 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="831dbf28-be96-4229-bd44-f88bb5b34e9c" containerName="neutron-sriov-openstack-openstack-cell1" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.936785 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.949270 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.949519 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.949745 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.949881 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.950060 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 09:09:31 crc kubenswrapper[4945]: I1206 09:09:31.956599 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-whzbj"] Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.017565 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctr6k\" (UniqueName: \"kubernetes.io/projected/94f979eb-bec9-48f7-b87f-f9618e999665-kube-api-access-ctr6k\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.017667 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.018221 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.018326 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.018698 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.018966 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.121228 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.121719 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.121778 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctr6k\" (UniqueName: \"kubernetes.io/projected/94f979eb-bec9-48f7-b87f-f9618e999665-kube-api-access-ctr6k\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.121809 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.121905 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.121932 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.126814 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.128043 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.129872 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.130559 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.140481 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.144726 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctr6k\" (UniqueName: \"kubernetes.io/projected/94f979eb-bec9-48f7-b87f-f9618e999665-kube-api-access-ctr6k\") pod \"neutron-dhcp-openstack-openstack-cell1-whzbj\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.270775 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.864801 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-whzbj"] Dec 06 09:09:32 crc kubenswrapper[4945]: I1206 09:09:32.868385 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 09:09:33 crc kubenswrapper[4945]: I1206 09:09:33.855031 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" event={"ID":"94f979eb-bec9-48f7-b87f-f9618e999665","Type":"ContainerStarted","Data":"34b0d6542e2b9a5db9b1e6374abe44b29f859ace2c6fd9b9f1a9720682bb4544"} Dec 06 09:09:33 crc kubenswrapper[4945]: I1206 09:09:33.855671 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" event={"ID":"94f979eb-bec9-48f7-b87f-f9618e999665","Type":"ContainerStarted","Data":"a48151a3ce0a13a6d3828a4fdb84d52f1fe0227cc90eebf209b3f351acb2de62"} Dec 06 09:09:37 crc kubenswrapper[4945]: I1206 09:09:37.953836 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:09:37 crc kubenswrapper[4945]: E1206 09:09:37.954706 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:09:48 crc kubenswrapper[4945]: I1206 09:09:48.954140 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:09:48 crc kubenswrapper[4945]: E1206 09:09:48.954973 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:10:02 crc kubenswrapper[4945]: I1206 09:10:02.953503 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:10:02 crc kubenswrapper[4945]: E1206 09:10:02.954669 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:10:14 crc kubenswrapper[4945]: I1206 09:10:14.953621 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:10:14 crc kubenswrapper[4945]: E1206 09:10:14.955479 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:10:25 crc kubenswrapper[4945]: I1206 09:10:25.953777 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:10:25 crc kubenswrapper[4945]: E1206 09:10:25.955030 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.228600 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" podStartSLOduration=55.5719474 podStartE2EDuration="56.228574819s" podCreationTimestamp="2025-12-06 09:09:31 +0000 UTC" firstStartedPulling="2025-12-06 09:09:32.86811633 +0000 UTC m=+8186.322977374" lastFinishedPulling="2025-12-06 09:09:33.524743749 +0000 UTC m=+8186.979604793" observedRunningTime="2025-12-06 09:09:33.881210382 +0000 UTC m=+8187.336071436" watchObservedRunningTime="2025-12-06 09:10:27.228574819 +0000 UTC m=+8240.683435863" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.230662 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xw4s2"] Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.233492 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.250432 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xw4s2"] Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.371212 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-utilities\") pod \"certified-operators-xw4s2\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.371588 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-catalog-content\") pod \"certified-operators-xw4s2\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.371757 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fs69\" (UniqueName: \"kubernetes.io/projected/747da355-bc41-480b-8e8d-49a8c0ef46fa-kube-api-access-6fs69\") pod \"certified-operators-xw4s2\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.473852 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fs69\" (UniqueName: \"kubernetes.io/projected/747da355-bc41-480b-8e8d-49a8c0ef46fa-kube-api-access-6fs69\") pod \"certified-operators-xw4s2\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.474039 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-utilities\") pod \"certified-operators-xw4s2\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.474108 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-catalog-content\") pod \"certified-operators-xw4s2\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.474748 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-catalog-content\") pod \"certified-operators-xw4s2\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.474818 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-utilities\") pod \"certified-operators-xw4s2\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.495093 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fs69\" (UniqueName: \"kubernetes.io/projected/747da355-bc41-480b-8e8d-49a8c0ef46fa-kube-api-access-6fs69\") pod \"certified-operators-xw4s2\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:27 crc kubenswrapper[4945]: I1206 09:10:27.568736 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:28 crc kubenswrapper[4945]: I1206 09:10:28.208070 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xw4s2"] Dec 06 09:10:28 crc kubenswrapper[4945]: W1206 09:10:28.211883 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod747da355_bc41_480b_8e8d_49a8c0ef46fa.slice/crio-dfa4d9fe0140df1cef4fb05f5c4b09bee02695c671db2f4adf72ca2a5b9e901c WatchSource:0}: Error finding container dfa4d9fe0140df1cef4fb05f5c4b09bee02695c671db2f4adf72ca2a5b9e901c: Status 404 returned error can't find the container with id dfa4d9fe0140df1cef4fb05f5c4b09bee02695c671db2f4adf72ca2a5b9e901c Dec 06 09:10:28 crc kubenswrapper[4945]: I1206 09:10:28.409791 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw4s2" event={"ID":"747da355-bc41-480b-8e8d-49a8c0ef46fa","Type":"ContainerStarted","Data":"dfa4d9fe0140df1cef4fb05f5c4b09bee02695c671db2f4adf72ca2a5b9e901c"} Dec 06 09:10:29 crc kubenswrapper[4945]: I1206 09:10:29.423031 4945 generic.go:334] "Generic (PLEG): container finished" podID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerID="e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5" exitCode=0 Dec 06 09:10:29 crc kubenswrapper[4945]: I1206 09:10:29.423294 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw4s2" event={"ID":"747da355-bc41-480b-8e8d-49a8c0ef46fa","Type":"ContainerDied","Data":"e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5"} Dec 06 09:10:30 crc kubenswrapper[4945]: I1206 09:10:30.450936 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw4s2" event={"ID":"747da355-bc41-480b-8e8d-49a8c0ef46fa","Type":"ContainerStarted","Data":"081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975"} Dec 06 09:10:31 crc kubenswrapper[4945]: I1206 09:10:31.462217 4945 generic.go:334] "Generic (PLEG): container finished" podID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerID="081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975" exitCode=0 Dec 06 09:10:31 crc kubenswrapper[4945]: I1206 09:10:31.462333 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw4s2" event={"ID":"747da355-bc41-480b-8e8d-49a8c0ef46fa","Type":"ContainerDied","Data":"081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975"} Dec 06 09:10:32 crc kubenswrapper[4945]: I1206 09:10:32.473420 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw4s2" event={"ID":"747da355-bc41-480b-8e8d-49a8c0ef46fa","Type":"ContainerStarted","Data":"604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228"} Dec 06 09:10:32 crc kubenswrapper[4945]: I1206 09:10:32.497127 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xw4s2" podStartSLOduration=3.10439429 podStartE2EDuration="5.497113542s" podCreationTimestamp="2025-12-06 09:10:27 +0000 UTC" firstStartedPulling="2025-12-06 09:10:29.436931972 +0000 UTC m=+8242.891793016" lastFinishedPulling="2025-12-06 09:10:31.829651224 +0000 UTC m=+8245.284512268" observedRunningTime="2025-12-06 09:10:32.489070487 +0000 UTC m=+8245.943931531" watchObservedRunningTime="2025-12-06 09:10:32.497113542 +0000 UTC m=+8245.951974586" Dec 06 09:10:37 crc kubenswrapper[4945]: I1206 09:10:37.569258 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:37 crc kubenswrapper[4945]: I1206 09:10:37.569868 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:37 crc kubenswrapper[4945]: I1206 09:10:37.614109 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:38 crc kubenswrapper[4945]: I1206 09:10:38.578370 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:38 crc kubenswrapper[4945]: I1206 09:10:38.627400 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xw4s2"] Dec 06 09:10:40 crc kubenswrapper[4945]: I1206 09:10:40.558846 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xw4s2" podUID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerName="registry-server" containerID="cri-o://604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228" gracePeriod=2 Dec 06 09:10:40 crc kubenswrapper[4945]: I1206 09:10:40.954681 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:10:40 crc kubenswrapper[4945]: E1206 09:10:40.955438 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.036571 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.075467 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-catalog-content\") pod \"747da355-bc41-480b-8e8d-49a8c0ef46fa\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.075620 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-utilities\") pod \"747da355-bc41-480b-8e8d-49a8c0ef46fa\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.075640 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fs69\" (UniqueName: \"kubernetes.io/projected/747da355-bc41-480b-8e8d-49a8c0ef46fa-kube-api-access-6fs69\") pod \"747da355-bc41-480b-8e8d-49a8c0ef46fa\" (UID: \"747da355-bc41-480b-8e8d-49a8c0ef46fa\") " Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.077509 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-utilities" (OuterVolumeSpecName: "utilities") pod "747da355-bc41-480b-8e8d-49a8c0ef46fa" (UID: "747da355-bc41-480b-8e8d-49a8c0ef46fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.082604 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/747da355-bc41-480b-8e8d-49a8c0ef46fa-kube-api-access-6fs69" (OuterVolumeSpecName: "kube-api-access-6fs69") pod "747da355-bc41-480b-8e8d-49a8c0ef46fa" (UID: "747da355-bc41-480b-8e8d-49a8c0ef46fa"). InnerVolumeSpecName "kube-api-access-6fs69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.124044 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "747da355-bc41-480b-8e8d-49a8c0ef46fa" (UID: "747da355-bc41-480b-8e8d-49a8c0ef46fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.178430 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.178748 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747da355-bc41-480b-8e8d-49a8c0ef46fa-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.178765 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fs69\" (UniqueName: \"kubernetes.io/projected/747da355-bc41-480b-8e8d-49a8c0ef46fa-kube-api-access-6fs69\") on node \"crc\" DevicePath \"\"" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.569851 4945 generic.go:334] "Generic (PLEG): container finished" podID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerID="604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228" exitCode=0 Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.569893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw4s2" event={"ID":"747da355-bc41-480b-8e8d-49a8c0ef46fa","Type":"ContainerDied","Data":"604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228"} Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.569920 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw4s2" event={"ID":"747da355-bc41-480b-8e8d-49a8c0ef46fa","Type":"ContainerDied","Data":"dfa4d9fe0140df1cef4fb05f5c4b09bee02695c671db2f4adf72ca2a5b9e901c"} Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.569925 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xw4s2" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.569940 4945 scope.go:117] "RemoveContainer" containerID="604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.590174 4945 scope.go:117] "RemoveContainer" containerID="081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.609193 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xw4s2"] Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.621058 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xw4s2"] Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.632903 4945 scope.go:117] "RemoveContainer" containerID="e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.670097 4945 scope.go:117] "RemoveContainer" containerID="604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228" Dec 06 09:10:41 crc kubenswrapper[4945]: E1206 09:10:41.670602 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228\": container with ID starting with 604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228 not found: ID does not exist" containerID="604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.670637 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228"} err="failed to get container status \"604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228\": rpc error: code = NotFound desc = could not find container \"604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228\": container with ID starting with 604c335a0226570b10e9876718865fa2eea8707c4e7bde7a22965bc020094228 not found: ID does not exist" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.670657 4945 scope.go:117] "RemoveContainer" containerID="081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975" Dec 06 09:10:41 crc kubenswrapper[4945]: E1206 09:10:41.671091 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975\": container with ID starting with 081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975 not found: ID does not exist" containerID="081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.671140 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975"} err="failed to get container status \"081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975\": rpc error: code = NotFound desc = could not find container \"081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975\": container with ID starting with 081cb653f46206a214741818dde03296b9309f39099e20e45a20a835b29d5975 not found: ID does not exist" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.671173 4945 scope.go:117] "RemoveContainer" containerID="e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5" Dec 06 09:10:41 crc kubenswrapper[4945]: E1206 09:10:41.671747 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5\": container with ID starting with e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5 not found: ID does not exist" containerID="e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5" Dec 06 09:10:41 crc kubenswrapper[4945]: I1206 09:10:41.671777 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5"} err="failed to get container status \"e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5\": rpc error: code = NotFound desc = could not find container \"e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5\": container with ID starting with e291889bfef537407b9c0f6e8ab6216a25830374b8c3d328e66cef7d34e4b0f5 not found: ID does not exist" Dec 06 09:10:42 crc kubenswrapper[4945]: I1206 09:10:42.967622 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="747da355-bc41-480b-8e8d-49a8c0ef46fa" path="/var/lib/kubelet/pods/747da355-bc41-480b-8e8d-49a8c0ef46fa/volumes" Dec 06 09:10:52 crc kubenswrapper[4945]: I1206 09:10:52.953144 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:10:52 crc kubenswrapper[4945]: E1206 09:10:52.954122 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:11:05 crc kubenswrapper[4945]: I1206 09:11:05.954202 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:11:05 crc kubenswrapper[4945]: E1206 09:11:05.955414 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:11:17 crc kubenswrapper[4945]: I1206 09:11:17.954240 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:11:17 crc kubenswrapper[4945]: E1206 09:11:17.955593 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:11:32 crc kubenswrapper[4945]: I1206 09:11:32.954779 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:11:32 crc kubenswrapper[4945]: E1206 09:11:32.956475 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:11:45 crc kubenswrapper[4945]: I1206 09:11:45.953778 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:11:45 crc kubenswrapper[4945]: E1206 09:11:45.954579 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:11:56 crc kubenswrapper[4945]: I1206 09:11:56.959377 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:11:56 crc kubenswrapper[4945]: E1206 09:11:56.960234 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:12:10 crc kubenswrapper[4945]: I1206 09:12:10.954685 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:12:10 crc kubenswrapper[4945]: E1206 09:12:10.955569 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:12:25 crc kubenswrapper[4945]: I1206 09:12:25.953799 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:12:25 crc kubenswrapper[4945]: E1206 09:12:25.954700 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:12:39 crc kubenswrapper[4945]: I1206 09:12:39.955645 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:12:39 crc kubenswrapper[4945]: E1206 09:12:39.956638 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:12:51 crc kubenswrapper[4945]: I1206 09:12:51.954187 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:12:51 crc kubenswrapper[4945]: E1206 09:12:51.957013 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:13:04 crc kubenswrapper[4945]: I1206 09:13:04.953435 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:13:04 crc kubenswrapper[4945]: E1206 09:13:04.954445 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:13:18 crc kubenswrapper[4945]: I1206 09:13:18.954867 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:13:18 crc kubenswrapper[4945]: E1206 09:13:18.959652 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:13:32 crc kubenswrapper[4945]: I1206 09:13:32.953309 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:13:32 crc kubenswrapper[4945]: E1206 09:13:32.954532 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.243207 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v5tfj"] Dec 06 09:13:33 crc kubenswrapper[4945]: E1206 09:13:33.243760 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerName="extract-utilities" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.243783 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerName="extract-utilities" Dec 06 09:13:33 crc kubenswrapper[4945]: E1206 09:13:33.243829 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerName="registry-server" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.243838 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerName="registry-server" Dec 06 09:13:33 crc kubenswrapper[4945]: E1206 09:13:33.243873 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerName="extract-content" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.243881 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerName="extract-content" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.244156 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="747da355-bc41-480b-8e8d-49a8c0ef46fa" containerName="registry-server" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.246221 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.271113 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5tfj"] Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.316264 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pcrd\" (UniqueName: \"kubernetes.io/projected/dc4d0c15-10dc-47ee-9d0d-8d990849a464-kube-api-access-2pcrd\") pod \"redhat-marketplace-v5tfj\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.316334 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-catalog-content\") pod \"redhat-marketplace-v5tfj\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.316417 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-utilities\") pod \"redhat-marketplace-v5tfj\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.418061 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pcrd\" (UniqueName: \"kubernetes.io/projected/dc4d0c15-10dc-47ee-9d0d-8d990849a464-kube-api-access-2pcrd\") pod \"redhat-marketplace-v5tfj\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.418103 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-catalog-content\") pod \"redhat-marketplace-v5tfj\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.418138 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-utilities\") pod \"redhat-marketplace-v5tfj\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.418790 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-catalog-content\") pod \"redhat-marketplace-v5tfj\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.418816 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-utilities\") pod \"redhat-marketplace-v5tfj\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.450777 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pcrd\" (UniqueName: \"kubernetes.io/projected/dc4d0c15-10dc-47ee-9d0d-8d990849a464-kube-api-access-2pcrd\") pod \"redhat-marketplace-v5tfj\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.451051 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p6pft"] Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.456795 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.498483 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p6pft"] Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.615916 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.624583 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-utilities\") pod \"redhat-operators-p6pft\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.624663 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwrxf\" (UniqueName: \"kubernetes.io/projected/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-kube-api-access-lwrxf\") pod \"redhat-operators-p6pft\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.624870 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-catalog-content\") pod \"redhat-operators-p6pft\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.727018 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-utilities\") pod \"redhat-operators-p6pft\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.727099 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwrxf\" (UniqueName: \"kubernetes.io/projected/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-kube-api-access-lwrxf\") pod \"redhat-operators-p6pft\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.727251 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-catalog-content\") pod \"redhat-operators-p6pft\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.727771 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-utilities\") pod \"redhat-operators-p6pft\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.727865 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-catalog-content\") pod \"redhat-operators-p6pft\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.758507 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwrxf\" (UniqueName: \"kubernetes.io/projected/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-kube-api-access-lwrxf\") pod \"redhat-operators-p6pft\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:33 crc kubenswrapper[4945]: I1206 09:13:33.833891 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:34 crc kubenswrapper[4945]: I1206 09:13:34.095375 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5tfj"] Dec 06 09:13:34 crc kubenswrapper[4945]: I1206 09:13:34.321616 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p6pft"] Dec 06 09:13:34 crc kubenswrapper[4945]: W1206 09:13:34.366438 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2203e77f_dcd0_4cb6_a8eb_6b31c22105ae.slice/crio-bdb104b07a3cf7a37050fcd830f019263268fdb86599ed29a263158468530454 WatchSource:0}: Error finding container bdb104b07a3cf7a37050fcd830f019263268fdb86599ed29a263158468530454: Status 404 returned error can't find the container with id bdb104b07a3cf7a37050fcd830f019263268fdb86599ed29a263158468530454 Dec 06 09:13:34 crc kubenswrapper[4945]: I1206 09:13:34.676518 4945 generic.go:334] "Generic (PLEG): container finished" podID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerID="5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e" exitCode=0 Dec 06 09:13:34 crc kubenswrapper[4945]: I1206 09:13:34.676731 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6pft" event={"ID":"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae","Type":"ContainerDied","Data":"5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e"} Dec 06 09:13:34 crc kubenswrapper[4945]: I1206 09:13:34.676852 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6pft" event={"ID":"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae","Type":"ContainerStarted","Data":"bdb104b07a3cf7a37050fcd830f019263268fdb86599ed29a263158468530454"} Dec 06 09:13:34 crc kubenswrapper[4945]: I1206 09:13:34.691469 4945 generic.go:334] "Generic (PLEG): container finished" podID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerID="990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308" exitCode=0 Dec 06 09:13:34 crc kubenswrapper[4945]: I1206 09:13:34.691514 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5tfj" event={"ID":"dc4d0c15-10dc-47ee-9d0d-8d990849a464","Type":"ContainerDied","Data":"990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308"} Dec 06 09:13:34 crc kubenswrapper[4945]: I1206 09:13:34.691555 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5tfj" event={"ID":"dc4d0c15-10dc-47ee-9d0d-8d990849a464","Type":"ContainerStarted","Data":"926eb004111dd0b7d7e978fa72b7080e786a2f03eb77defe66f24a8243d60128"} Dec 06 09:13:35 crc kubenswrapper[4945]: I1206 09:13:35.702404 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5tfj" event={"ID":"dc4d0c15-10dc-47ee-9d0d-8d990849a464","Type":"ContainerStarted","Data":"2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c"} Dec 06 09:13:35 crc kubenswrapper[4945]: I1206 09:13:35.705160 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6pft" event={"ID":"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae","Type":"ContainerStarted","Data":"b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0"} Dec 06 09:13:36 crc kubenswrapper[4945]: I1206 09:13:36.726854 4945 generic.go:334] "Generic (PLEG): container finished" podID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerID="2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c" exitCode=0 Dec 06 09:13:36 crc kubenswrapper[4945]: I1206 09:13:36.726966 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5tfj" event={"ID":"dc4d0c15-10dc-47ee-9d0d-8d990849a464","Type":"ContainerDied","Data":"2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c"} Dec 06 09:13:40 crc kubenswrapper[4945]: I1206 09:13:40.773604 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5tfj" event={"ID":"dc4d0c15-10dc-47ee-9d0d-8d990849a464","Type":"ContainerStarted","Data":"929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315"} Dec 06 09:13:40 crc kubenswrapper[4945]: I1206 09:13:40.776190 4945 generic.go:334] "Generic (PLEG): container finished" podID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerID="b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0" exitCode=0 Dec 06 09:13:40 crc kubenswrapper[4945]: I1206 09:13:40.776225 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6pft" event={"ID":"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae","Type":"ContainerDied","Data":"b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0"} Dec 06 09:13:40 crc kubenswrapper[4945]: I1206 09:13:40.813765 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v5tfj" podStartSLOduration=2.052081704 podStartE2EDuration="7.813745702s" podCreationTimestamp="2025-12-06 09:13:33 +0000 UTC" firstStartedPulling="2025-12-06 09:13:34.69612998 +0000 UTC m=+8428.150991024" lastFinishedPulling="2025-12-06 09:13:40.457793988 +0000 UTC m=+8433.912655022" observedRunningTime="2025-12-06 09:13:40.804464263 +0000 UTC m=+8434.259325307" watchObservedRunningTime="2025-12-06 09:13:40.813745702 +0000 UTC m=+8434.268606746" Dec 06 09:13:41 crc kubenswrapper[4945]: I1206 09:13:41.803324 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6pft" event={"ID":"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae","Type":"ContainerStarted","Data":"d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318"} Dec 06 09:13:41 crc kubenswrapper[4945]: I1206 09:13:41.827835 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p6pft" podStartSLOduration=2.341225022 podStartE2EDuration="8.827817456s" podCreationTimestamp="2025-12-06 09:13:33 +0000 UTC" firstStartedPulling="2025-12-06 09:13:34.681233052 +0000 UTC m=+8428.136094096" lastFinishedPulling="2025-12-06 09:13:41.167825496 +0000 UTC m=+8434.622686530" observedRunningTime="2025-12-06 09:13:41.820353386 +0000 UTC m=+8435.275214440" watchObservedRunningTime="2025-12-06 09:13:41.827817456 +0000 UTC m=+8435.282678500" Dec 06 09:13:43 crc kubenswrapper[4945]: I1206 09:13:43.616371 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:43 crc kubenswrapper[4945]: I1206 09:13:43.616762 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:43 crc kubenswrapper[4945]: I1206 09:13:43.697376 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:43 crc kubenswrapper[4945]: I1206 09:13:43.834847 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:43 crc kubenswrapper[4945]: I1206 09:13:43.835154 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:44 crc kubenswrapper[4945]: I1206 09:13:44.879932 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p6pft" podUID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerName="registry-server" probeResult="failure" output=< Dec 06 09:13:44 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 09:13:44 crc kubenswrapper[4945]: > Dec 06 09:13:47 crc kubenswrapper[4945]: I1206 09:13:47.957872 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:13:47 crc kubenswrapper[4945]: E1206 09:13:47.958509 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:13:53 crc kubenswrapper[4945]: I1206 09:13:53.676061 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:53 crc kubenswrapper[4945]: I1206 09:13:53.739093 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5tfj"] Dec 06 09:13:53 crc kubenswrapper[4945]: I1206 09:13:53.893606 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:53 crc kubenswrapper[4945]: I1206 09:13:53.925774 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v5tfj" podUID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerName="registry-server" containerID="cri-o://929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315" gracePeriod=2 Dec 06 09:13:53 crc kubenswrapper[4945]: I1206 09:13:53.943148 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.399765 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.502018 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pcrd\" (UniqueName: \"kubernetes.io/projected/dc4d0c15-10dc-47ee-9d0d-8d990849a464-kube-api-access-2pcrd\") pod \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.502088 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-catalog-content\") pod \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.502135 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-utilities\") pod \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\" (UID: \"dc4d0c15-10dc-47ee-9d0d-8d990849a464\") " Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.502936 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-utilities" (OuterVolumeSpecName: "utilities") pod "dc4d0c15-10dc-47ee-9d0d-8d990849a464" (UID: "dc4d0c15-10dc-47ee-9d0d-8d990849a464"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.512462 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc4d0c15-10dc-47ee-9d0d-8d990849a464-kube-api-access-2pcrd" (OuterVolumeSpecName: "kube-api-access-2pcrd") pod "dc4d0c15-10dc-47ee-9d0d-8d990849a464" (UID: "dc4d0c15-10dc-47ee-9d0d-8d990849a464"). InnerVolumeSpecName "kube-api-access-2pcrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.521605 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc4d0c15-10dc-47ee-9d0d-8d990849a464" (UID: "dc4d0c15-10dc-47ee-9d0d-8d990849a464"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.604452 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pcrd\" (UniqueName: \"kubernetes.io/projected/dc4d0c15-10dc-47ee-9d0d-8d990849a464-kube-api-access-2pcrd\") on node \"crc\" DevicePath \"\"" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.604488 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.604504 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4d0c15-10dc-47ee-9d0d-8d990849a464-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.937354 4945 generic.go:334] "Generic (PLEG): container finished" podID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerID="929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315" exitCode=0 Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.937414 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5tfj" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.937408 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5tfj" event={"ID":"dc4d0c15-10dc-47ee-9d0d-8d990849a464","Type":"ContainerDied","Data":"929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315"} Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.937492 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5tfj" event={"ID":"dc4d0c15-10dc-47ee-9d0d-8d990849a464","Type":"ContainerDied","Data":"926eb004111dd0b7d7e978fa72b7080e786a2f03eb77defe66f24a8243d60128"} Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.937511 4945 scope.go:117] "RemoveContainer" containerID="929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.971926 4945 scope.go:117] "RemoveContainer" containerID="2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c" Dec 06 09:13:54 crc kubenswrapper[4945]: I1206 09:13:54.994976 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5tfj"] Dec 06 09:13:55 crc kubenswrapper[4945]: I1206 09:13:55.001124 4945 scope.go:117] "RemoveContainer" containerID="990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308" Dec 06 09:13:55 crc kubenswrapper[4945]: I1206 09:13:55.005224 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5tfj"] Dec 06 09:13:55 crc kubenswrapper[4945]: I1206 09:13:55.044900 4945 scope.go:117] "RemoveContainer" containerID="929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315" Dec 06 09:13:55 crc kubenswrapper[4945]: E1206 09:13:55.045635 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315\": container with ID starting with 929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315 not found: ID does not exist" containerID="929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315" Dec 06 09:13:55 crc kubenswrapper[4945]: I1206 09:13:55.045704 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315"} err="failed to get container status \"929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315\": rpc error: code = NotFound desc = could not find container \"929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315\": container with ID starting with 929945c71dc845591fb83fe421f665edf501b8871935fe163e32ab96912a9315 not found: ID does not exist" Dec 06 09:13:55 crc kubenswrapper[4945]: I1206 09:13:55.045740 4945 scope.go:117] "RemoveContainer" containerID="2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c" Dec 06 09:13:55 crc kubenswrapper[4945]: E1206 09:13:55.046039 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c\": container with ID starting with 2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c not found: ID does not exist" containerID="2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c" Dec 06 09:13:55 crc kubenswrapper[4945]: I1206 09:13:55.046078 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c"} err="failed to get container status \"2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c\": rpc error: code = NotFound desc = could not find container \"2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c\": container with ID starting with 2964fffd3ea5663e02988aebe7e702f0cbda752e959c8b737aef5062a14a437c not found: ID does not exist" Dec 06 09:13:55 crc kubenswrapper[4945]: I1206 09:13:55.046097 4945 scope.go:117] "RemoveContainer" containerID="990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308" Dec 06 09:13:55 crc kubenswrapper[4945]: E1206 09:13:55.046549 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308\": container with ID starting with 990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308 not found: ID does not exist" containerID="990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308" Dec 06 09:13:55 crc kubenswrapper[4945]: I1206 09:13:55.046580 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308"} err="failed to get container status \"990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308\": rpc error: code = NotFound desc = could not find container \"990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308\": container with ID starting with 990bb6754e7a7ab58c9ed17458ccc08fdad73478f80e5f0630432b54dc098308 not found: ID does not exist" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.121158 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p6pft"] Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.121513 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p6pft" podUID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerName="registry-server" containerID="cri-o://d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318" gracePeriod=2 Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.570926 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.645597 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwrxf\" (UniqueName: \"kubernetes.io/projected/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-kube-api-access-lwrxf\") pod \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.645772 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-utilities\") pod \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.645848 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-catalog-content\") pod \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\" (UID: \"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae\") " Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.646828 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-utilities" (OuterVolumeSpecName: "utilities") pod "2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" (UID: "2203e77f-dcd0-4cb6-a8eb-6b31c22105ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.652379 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-kube-api-access-lwrxf" (OuterVolumeSpecName: "kube-api-access-lwrxf") pod "2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" (UID: "2203e77f-dcd0-4cb6-a8eb-6b31c22105ae"). InnerVolumeSpecName "kube-api-access-lwrxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.742401 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" (UID: "2203e77f-dcd0-4cb6-a8eb-6b31c22105ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.749204 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwrxf\" (UniqueName: \"kubernetes.io/projected/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-kube-api-access-lwrxf\") on node \"crc\" DevicePath \"\"" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.749255 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.749269 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.961496 4945 generic.go:334] "Generic (PLEG): container finished" podID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerID="d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318" exitCode=0 Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.963180 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p6pft" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.964093 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" path="/var/lib/kubelet/pods/dc4d0c15-10dc-47ee-9d0d-8d990849a464/volumes" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.964884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6pft" event={"ID":"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae","Type":"ContainerDied","Data":"d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318"} Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.964918 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6pft" event={"ID":"2203e77f-dcd0-4cb6-a8eb-6b31c22105ae","Type":"ContainerDied","Data":"bdb104b07a3cf7a37050fcd830f019263268fdb86599ed29a263158468530454"} Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.964948 4945 scope.go:117] "RemoveContainer" containerID="d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318" Dec 06 09:13:56 crc kubenswrapper[4945]: I1206 09:13:56.986914 4945 scope.go:117] "RemoveContainer" containerID="b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0" Dec 06 09:13:57 crc kubenswrapper[4945]: I1206 09:13:57.008936 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p6pft"] Dec 06 09:13:57 crc kubenswrapper[4945]: I1206 09:13:57.014762 4945 scope.go:117] "RemoveContainer" containerID="5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e" Dec 06 09:13:57 crc kubenswrapper[4945]: I1206 09:13:57.018238 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p6pft"] Dec 06 09:13:57 crc kubenswrapper[4945]: I1206 09:13:57.067608 4945 scope.go:117] "RemoveContainer" containerID="d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318" Dec 06 09:13:57 crc kubenswrapper[4945]: E1206 09:13:57.068874 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318\": container with ID starting with d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318 not found: ID does not exist" containerID="d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318" Dec 06 09:13:57 crc kubenswrapper[4945]: I1206 09:13:57.068918 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318"} err="failed to get container status \"d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318\": rpc error: code = NotFound desc = could not find container \"d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318\": container with ID starting with d7e30ad7e3b2999f2b07e40e77996bef2ecab1d40d549887053d00206b5de318 not found: ID does not exist" Dec 06 09:13:57 crc kubenswrapper[4945]: I1206 09:13:57.068943 4945 scope.go:117] "RemoveContainer" containerID="b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0" Dec 06 09:13:57 crc kubenswrapper[4945]: E1206 09:13:57.070315 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0\": container with ID starting with b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0 not found: ID does not exist" containerID="b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0" Dec 06 09:13:57 crc kubenswrapper[4945]: I1206 09:13:57.070344 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0"} err="failed to get container status \"b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0\": rpc error: code = NotFound desc = could not find container \"b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0\": container with ID starting with b25231ed4fe6d5c9695461a580283fafff6c8c61b09a381f792ea910e3ebf5e0 not found: ID does not exist" Dec 06 09:13:57 crc kubenswrapper[4945]: I1206 09:13:57.070361 4945 scope.go:117] "RemoveContainer" containerID="5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e" Dec 06 09:13:57 crc kubenswrapper[4945]: E1206 09:13:57.070652 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e\": container with ID starting with 5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e not found: ID does not exist" containerID="5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e" Dec 06 09:13:57 crc kubenswrapper[4945]: I1206 09:13:57.070695 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e"} err="failed to get container status \"5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e\": rpc error: code = NotFound desc = could not find container \"5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e\": container with ID starting with 5549e2d1ef1b4168ceeb73fc5b5873dfb78c71e38f73cc5b162ce6c2deafba6e not found: ID does not exist" Dec 06 09:13:58 crc kubenswrapper[4945]: I1206 09:13:58.969048 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" path="/var/lib/kubelet/pods/2203e77f-dcd0-4cb6-a8eb-6b31c22105ae/volumes" Dec 06 09:13:59 crc kubenswrapper[4945]: I1206 09:13:59.954689 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:13:59 crc kubenswrapper[4945]: E1206 09:13:59.955601 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:14:13 crc kubenswrapper[4945]: I1206 09:14:13.954393 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:14:15 crc kubenswrapper[4945]: I1206 09:14:15.164141 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"6971dd6d12e521980a0aacdcb6dd8e0a16a4b48aeec4b9057f6e6d63ff4f22c6"} Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.189091 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f"] Dec 06 09:15:00 crc kubenswrapper[4945]: E1206 09:15:00.190163 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerName="registry-server" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.190180 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerName="registry-server" Dec 06 09:15:00 crc kubenswrapper[4945]: E1206 09:15:00.190203 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerName="extract-utilities" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.190209 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerName="extract-utilities" Dec 06 09:15:00 crc kubenswrapper[4945]: E1206 09:15:00.190222 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerName="extract-content" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.190229 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerName="extract-content" Dec 06 09:15:00 crc kubenswrapper[4945]: E1206 09:15:00.190246 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerName="extract-utilities" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.190252 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerName="extract-utilities" Dec 06 09:15:00 crc kubenswrapper[4945]: E1206 09:15:00.190272 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerName="registry-server" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.190296 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerName="registry-server" Dec 06 09:15:00 crc kubenswrapper[4945]: E1206 09:15:00.190308 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerName="extract-content" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.190314 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerName="extract-content" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.190513 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4d0c15-10dc-47ee-9d0d-8d990849a464" containerName="registry-server" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.190537 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2203e77f-dcd0-4cb6-a8eb-6b31c22105ae" containerName="registry-server" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.191339 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.195880 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.196749 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.219825 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f"] Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.391604 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8c05dd2-e91b-4464-916f-89507d2000e0-config-volume\") pod \"collect-profiles-29416875-5l99f\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.391672 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdpv6\" (UniqueName: \"kubernetes.io/projected/c8c05dd2-e91b-4464-916f-89507d2000e0-kube-api-access-jdpv6\") pod \"collect-profiles-29416875-5l99f\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.391988 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8c05dd2-e91b-4464-916f-89507d2000e0-secret-volume\") pod \"collect-profiles-29416875-5l99f\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.492977 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8c05dd2-e91b-4464-916f-89507d2000e0-secret-volume\") pod \"collect-profiles-29416875-5l99f\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.493055 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8c05dd2-e91b-4464-916f-89507d2000e0-config-volume\") pod \"collect-profiles-29416875-5l99f\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.493083 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdpv6\" (UniqueName: \"kubernetes.io/projected/c8c05dd2-e91b-4464-916f-89507d2000e0-kube-api-access-jdpv6\") pod \"collect-profiles-29416875-5l99f\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.494168 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8c05dd2-e91b-4464-916f-89507d2000e0-config-volume\") pod \"collect-profiles-29416875-5l99f\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.499522 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8c05dd2-e91b-4464-916f-89507d2000e0-secret-volume\") pod \"collect-profiles-29416875-5l99f\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.514170 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdpv6\" (UniqueName: \"kubernetes.io/projected/c8c05dd2-e91b-4464-916f-89507d2000e0-kube-api-access-jdpv6\") pod \"collect-profiles-29416875-5l99f\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:00 crc kubenswrapper[4945]: I1206 09:15:00.521540 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:01 crc kubenswrapper[4945]: I1206 09:15:01.017856 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f"] Dec 06 09:15:02 crc kubenswrapper[4945]: I1206 09:15:02.023258 4945 generic.go:334] "Generic (PLEG): container finished" podID="c8c05dd2-e91b-4464-916f-89507d2000e0" containerID="f9b789ca4fe4ab9c0fe5cb4735919a6e8af4df3a66cbb44c5b86ecab592c2b71" exitCode=0 Dec 06 09:15:02 crc kubenswrapper[4945]: I1206 09:15:02.023335 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" event={"ID":"c8c05dd2-e91b-4464-916f-89507d2000e0","Type":"ContainerDied","Data":"f9b789ca4fe4ab9c0fe5cb4735919a6e8af4df3a66cbb44c5b86ecab592c2b71"} Dec 06 09:15:02 crc kubenswrapper[4945]: I1206 09:15:02.023642 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" event={"ID":"c8c05dd2-e91b-4464-916f-89507d2000e0","Type":"ContainerStarted","Data":"0c02bbb5fbbd71714923637dbe75276bf3c41c707649f504afd4cde2f0a2948b"} Dec 06 09:15:03 crc kubenswrapper[4945]: I1206 09:15:03.388639 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:03 crc kubenswrapper[4945]: I1206 09:15:03.555900 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdpv6\" (UniqueName: \"kubernetes.io/projected/c8c05dd2-e91b-4464-916f-89507d2000e0-kube-api-access-jdpv6\") pod \"c8c05dd2-e91b-4464-916f-89507d2000e0\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " Dec 06 09:15:03 crc kubenswrapper[4945]: I1206 09:15:03.556029 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8c05dd2-e91b-4464-916f-89507d2000e0-config-volume\") pod \"c8c05dd2-e91b-4464-916f-89507d2000e0\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " Dec 06 09:15:03 crc kubenswrapper[4945]: I1206 09:15:03.556139 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8c05dd2-e91b-4464-916f-89507d2000e0-secret-volume\") pod \"c8c05dd2-e91b-4464-916f-89507d2000e0\" (UID: \"c8c05dd2-e91b-4464-916f-89507d2000e0\") " Dec 06 09:15:03 crc kubenswrapper[4945]: I1206 09:15:03.556538 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8c05dd2-e91b-4464-916f-89507d2000e0-config-volume" (OuterVolumeSpecName: "config-volume") pod "c8c05dd2-e91b-4464-916f-89507d2000e0" (UID: "c8c05dd2-e91b-4464-916f-89507d2000e0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 09:15:03 crc kubenswrapper[4945]: I1206 09:15:03.556792 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8c05dd2-e91b-4464-916f-89507d2000e0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 09:15:03 crc kubenswrapper[4945]: I1206 09:15:03.560924 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c05dd2-e91b-4464-916f-89507d2000e0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c8c05dd2-e91b-4464-916f-89507d2000e0" (UID: "c8c05dd2-e91b-4464-916f-89507d2000e0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:15:03 crc kubenswrapper[4945]: I1206 09:15:03.561465 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c05dd2-e91b-4464-916f-89507d2000e0-kube-api-access-jdpv6" (OuterVolumeSpecName: "kube-api-access-jdpv6") pod "c8c05dd2-e91b-4464-916f-89507d2000e0" (UID: "c8c05dd2-e91b-4464-916f-89507d2000e0"). InnerVolumeSpecName "kube-api-access-jdpv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:15:03 crc kubenswrapper[4945]: I1206 09:15:03.659076 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8c05dd2-e91b-4464-916f-89507d2000e0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 09:15:03 crc kubenswrapper[4945]: I1206 09:15:03.659111 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdpv6\" (UniqueName: \"kubernetes.io/projected/c8c05dd2-e91b-4464-916f-89507d2000e0-kube-api-access-jdpv6\") on node \"crc\" DevicePath \"\"" Dec 06 09:15:04 crc kubenswrapper[4945]: I1206 09:15:04.042664 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" event={"ID":"c8c05dd2-e91b-4464-916f-89507d2000e0","Type":"ContainerDied","Data":"0c02bbb5fbbd71714923637dbe75276bf3c41c707649f504afd4cde2f0a2948b"} Dec 06 09:15:04 crc kubenswrapper[4945]: I1206 09:15:04.042709 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c02bbb5fbbd71714923637dbe75276bf3c41c707649f504afd4cde2f0a2948b" Dec 06 09:15:04 crc kubenswrapper[4945]: I1206 09:15:04.042739 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416875-5l99f" Dec 06 09:15:04 crc kubenswrapper[4945]: I1206 09:15:04.466248 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz"] Dec 06 09:15:04 crc kubenswrapper[4945]: I1206 09:15:04.476585 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416830-sc8dz"] Dec 06 09:15:04 crc kubenswrapper[4945]: I1206 09:15:04.966788 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b16f13b-4a09-4bad-b647-655d12bd88c4" path="/var/lib/kubelet/pods/4b16f13b-4a09-4bad-b647-655d12bd88c4/volumes" Dec 06 09:15:29 crc kubenswrapper[4945]: I1206 09:15:29.631848 4945 scope.go:117] "RemoveContainer" containerID="3a19e0175f87a94f5da4ecb02686c5b6de18e7178d44591094410dbc86ca6bf7" Dec 06 09:15:48 crc kubenswrapper[4945]: I1206 09:15:48.903627 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cstq8"] Dec 06 09:15:48 crc kubenswrapper[4945]: E1206 09:15:48.905876 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c05dd2-e91b-4464-916f-89507d2000e0" containerName="collect-profiles" Dec 06 09:15:48 crc kubenswrapper[4945]: I1206 09:15:48.906019 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c05dd2-e91b-4464-916f-89507d2000e0" containerName="collect-profiles" Dec 06 09:15:48 crc kubenswrapper[4945]: I1206 09:15:48.906608 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c05dd2-e91b-4464-916f-89507d2000e0" containerName="collect-profiles" Dec 06 09:15:48 crc kubenswrapper[4945]: I1206 09:15:48.908273 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:48 crc kubenswrapper[4945]: I1206 09:15:48.913482 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cstq8"] Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.041425 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-catalog-content\") pod \"community-operators-cstq8\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.041556 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-utilities\") pod \"community-operators-cstq8\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.041759 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbhbb\" (UniqueName: \"kubernetes.io/projected/1b685769-099a-4def-a772-82adbde350b1-kube-api-access-gbhbb\") pod \"community-operators-cstq8\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.144092 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-catalog-content\") pod \"community-operators-cstq8\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.144207 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-utilities\") pod \"community-operators-cstq8\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.144264 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbhbb\" (UniqueName: \"kubernetes.io/projected/1b685769-099a-4def-a772-82adbde350b1-kube-api-access-gbhbb\") pod \"community-operators-cstq8\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.144814 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-catalog-content\") pod \"community-operators-cstq8\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.145073 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-utilities\") pod \"community-operators-cstq8\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.166243 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbhbb\" (UniqueName: \"kubernetes.io/projected/1b685769-099a-4def-a772-82adbde350b1-kube-api-access-gbhbb\") pod \"community-operators-cstq8\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.229341 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:49 crc kubenswrapper[4945]: I1206 09:15:49.862926 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cstq8"] Dec 06 09:15:50 crc kubenswrapper[4945]: I1206 09:15:50.549043 4945 generic.go:334] "Generic (PLEG): container finished" podID="1b685769-099a-4def-a772-82adbde350b1" containerID="e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded" exitCode=0 Dec 06 09:15:50 crc kubenswrapper[4945]: I1206 09:15:50.549311 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cstq8" event={"ID":"1b685769-099a-4def-a772-82adbde350b1","Type":"ContainerDied","Data":"e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded"} Dec 06 09:15:50 crc kubenswrapper[4945]: I1206 09:15:50.549415 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cstq8" event={"ID":"1b685769-099a-4def-a772-82adbde350b1","Type":"ContainerStarted","Data":"6a1e3528a0b03ca2c639e4511aa5fb6ba9ada15447070e0bd3c9665f96f4d29f"} Dec 06 09:15:50 crc kubenswrapper[4945]: I1206 09:15:50.551544 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 09:15:51 crc kubenswrapper[4945]: I1206 09:15:51.559881 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cstq8" event={"ID":"1b685769-099a-4def-a772-82adbde350b1","Type":"ContainerStarted","Data":"9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10"} Dec 06 09:15:52 crc kubenswrapper[4945]: I1206 09:15:52.570204 4945 generic.go:334] "Generic (PLEG): container finished" podID="1b685769-099a-4def-a772-82adbde350b1" containerID="9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10" exitCode=0 Dec 06 09:15:52 crc kubenswrapper[4945]: I1206 09:15:52.570264 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cstq8" event={"ID":"1b685769-099a-4def-a772-82adbde350b1","Type":"ContainerDied","Data":"9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10"} Dec 06 09:15:53 crc kubenswrapper[4945]: I1206 09:15:53.582004 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cstq8" event={"ID":"1b685769-099a-4def-a772-82adbde350b1","Type":"ContainerStarted","Data":"42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c"} Dec 06 09:15:53 crc kubenswrapper[4945]: I1206 09:15:53.612797 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cstq8" podStartSLOduration=3.199461182 podStartE2EDuration="5.612776775s" podCreationTimestamp="2025-12-06 09:15:48 +0000 UTC" firstStartedPulling="2025-12-06 09:15:50.551219315 +0000 UTC m=+8564.006080359" lastFinishedPulling="2025-12-06 09:15:52.964534908 +0000 UTC m=+8566.419395952" observedRunningTime="2025-12-06 09:15:53.601189365 +0000 UTC m=+8567.056050439" watchObservedRunningTime="2025-12-06 09:15:53.612776775 +0000 UTC m=+8567.067637839" Dec 06 09:15:59 crc kubenswrapper[4945]: I1206 09:15:59.230708 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:59 crc kubenswrapper[4945]: I1206 09:15:59.231251 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:59 crc kubenswrapper[4945]: I1206 09:15:59.302751 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:59 crc kubenswrapper[4945]: I1206 09:15:59.689980 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:15:59 crc kubenswrapper[4945]: I1206 09:15:59.740102 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cstq8"] Dec 06 09:16:01 crc kubenswrapper[4945]: I1206 09:16:01.663636 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cstq8" podUID="1b685769-099a-4def-a772-82adbde350b1" containerName="registry-server" containerID="cri-o://42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c" gracePeriod=2 Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.459904 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.522254 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-catalog-content\") pod \"1b685769-099a-4def-a772-82adbde350b1\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.522328 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbhbb\" (UniqueName: \"kubernetes.io/projected/1b685769-099a-4def-a772-82adbde350b1-kube-api-access-gbhbb\") pod \"1b685769-099a-4def-a772-82adbde350b1\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.522393 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-utilities\") pod \"1b685769-099a-4def-a772-82adbde350b1\" (UID: \"1b685769-099a-4def-a772-82adbde350b1\") " Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.523402 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-utilities" (OuterVolumeSpecName: "utilities") pod "1b685769-099a-4def-a772-82adbde350b1" (UID: "1b685769-099a-4def-a772-82adbde350b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.528419 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b685769-099a-4def-a772-82adbde350b1-kube-api-access-gbhbb" (OuterVolumeSpecName: "kube-api-access-gbhbb") pod "1b685769-099a-4def-a772-82adbde350b1" (UID: "1b685769-099a-4def-a772-82adbde350b1"). InnerVolumeSpecName "kube-api-access-gbhbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.573715 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b685769-099a-4def-a772-82adbde350b1" (UID: "1b685769-099a-4def-a772-82adbde350b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.624337 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.624368 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbhbb\" (UniqueName: \"kubernetes.io/projected/1b685769-099a-4def-a772-82adbde350b1-kube-api-access-gbhbb\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.624380 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b685769-099a-4def-a772-82adbde350b1-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.676144 4945 generic.go:334] "Generic (PLEG): container finished" podID="1b685769-099a-4def-a772-82adbde350b1" containerID="42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c" exitCode=0 Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.676194 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cstq8" event={"ID":"1b685769-099a-4def-a772-82adbde350b1","Type":"ContainerDied","Data":"42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c"} Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.676211 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cstq8" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.676231 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cstq8" event={"ID":"1b685769-099a-4def-a772-82adbde350b1","Type":"ContainerDied","Data":"6a1e3528a0b03ca2c639e4511aa5fb6ba9ada15447070e0bd3c9665f96f4d29f"} Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.676335 4945 scope.go:117] "RemoveContainer" containerID="42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.711572 4945 scope.go:117] "RemoveContainer" containerID="9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.720426 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cstq8"] Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.737994 4945 scope.go:117] "RemoveContainer" containerID="e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.750159 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cstq8"] Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.785017 4945 scope.go:117] "RemoveContainer" containerID="42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c" Dec 06 09:16:02 crc kubenswrapper[4945]: E1206 09:16:02.785514 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c\": container with ID starting with 42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c not found: ID does not exist" containerID="42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.785554 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c"} err="failed to get container status \"42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c\": rpc error: code = NotFound desc = could not find container \"42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c\": container with ID starting with 42ed2b439b662f47826a40af6ddfa1991646b97193bb468b386fae011e8df49c not found: ID does not exist" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.785581 4945 scope.go:117] "RemoveContainer" containerID="9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10" Dec 06 09:16:02 crc kubenswrapper[4945]: E1206 09:16:02.785820 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10\": container with ID starting with 9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10 not found: ID does not exist" containerID="9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.785845 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10"} err="failed to get container status \"9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10\": rpc error: code = NotFound desc = could not find container \"9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10\": container with ID starting with 9174b2b226426f088e721d474f5a86730d27a801a9091da684ab49f1dc332d10 not found: ID does not exist" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.785861 4945 scope.go:117] "RemoveContainer" containerID="e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded" Dec 06 09:16:02 crc kubenswrapper[4945]: E1206 09:16:02.786062 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded\": container with ID starting with e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded not found: ID does not exist" containerID="e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.786085 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded"} err="failed to get container status \"e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded\": rpc error: code = NotFound desc = could not find container \"e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded\": container with ID starting with e6997f1e8221d5d142a029ae46a81e86d057637fdbd145e5b29a36ccfea59ded not found: ID does not exist" Dec 06 09:16:02 crc kubenswrapper[4945]: I1206 09:16:02.964191 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b685769-099a-4def-a772-82adbde350b1" path="/var/lib/kubelet/pods/1b685769-099a-4def-a772-82adbde350b1/volumes" Dec 06 09:16:12 crc kubenswrapper[4945]: I1206 09:16:12.784197 4945 generic.go:334] "Generic (PLEG): container finished" podID="94f979eb-bec9-48f7-b87f-f9618e999665" containerID="34b0d6542e2b9a5db9b1e6374abe44b29f859ace2c6fd9b9f1a9720682bb4544" exitCode=0 Dec 06 09:16:12 crc kubenswrapper[4945]: I1206 09:16:12.784257 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" event={"ID":"94f979eb-bec9-48f7-b87f-f9618e999665","Type":"ContainerDied","Data":"34b0d6542e2b9a5db9b1e6374abe44b29f859ace2c6fd9b9f1a9720682bb4544"} Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.331647 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.363229 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-combined-ca-bundle\") pod \"94f979eb-bec9-48f7-b87f-f9618e999665\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.363340 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-agent-neutron-config-0\") pod \"94f979eb-bec9-48f7-b87f-f9618e999665\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.363399 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ssh-key\") pod \"94f979eb-bec9-48f7-b87f-f9618e999665\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.363460 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ceph\") pod \"94f979eb-bec9-48f7-b87f-f9618e999665\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.363508 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-inventory\") pod \"94f979eb-bec9-48f7-b87f-f9618e999665\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.363536 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctr6k\" (UniqueName: \"kubernetes.io/projected/94f979eb-bec9-48f7-b87f-f9618e999665-kube-api-access-ctr6k\") pod \"94f979eb-bec9-48f7-b87f-f9618e999665\" (UID: \"94f979eb-bec9-48f7-b87f-f9618e999665\") " Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.369740 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "94f979eb-bec9-48f7-b87f-f9618e999665" (UID: "94f979eb-bec9-48f7-b87f-f9618e999665"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.369886 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ceph" (OuterVolumeSpecName: "ceph") pod "94f979eb-bec9-48f7-b87f-f9618e999665" (UID: "94f979eb-bec9-48f7-b87f-f9618e999665"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.373620 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94f979eb-bec9-48f7-b87f-f9618e999665-kube-api-access-ctr6k" (OuterVolumeSpecName: "kube-api-access-ctr6k") pod "94f979eb-bec9-48f7-b87f-f9618e999665" (UID: "94f979eb-bec9-48f7-b87f-f9618e999665"). InnerVolumeSpecName "kube-api-access-ctr6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.396568 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "94f979eb-bec9-48f7-b87f-f9618e999665" (UID: "94f979eb-bec9-48f7-b87f-f9618e999665"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.397774 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-inventory" (OuterVolumeSpecName: "inventory") pod "94f979eb-bec9-48f7-b87f-f9618e999665" (UID: "94f979eb-bec9-48f7-b87f-f9618e999665"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.400048 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "94f979eb-bec9-48f7-b87f-f9618e999665" (UID: "94f979eb-bec9-48f7-b87f-f9618e999665"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.466450 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.466476 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.466487 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.466496 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.466506 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94f979eb-bec9-48f7-b87f-f9618e999665-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.466518 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctr6k\" (UniqueName: \"kubernetes.io/projected/94f979eb-bec9-48f7-b87f-f9618e999665-kube-api-access-ctr6k\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.805832 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" event={"ID":"94f979eb-bec9-48f7-b87f-f9618e999665","Type":"ContainerDied","Data":"a48151a3ce0a13a6d3828a4fdb84d52f1fe0227cc90eebf209b3f351acb2de62"} Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.805876 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-whzbj" Dec 06 09:16:14 crc kubenswrapper[4945]: I1206 09:16:14.805891 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a48151a3ce0a13a6d3828a4fdb84d52f1fe0227cc90eebf209b3f351acb2de62" Dec 06 09:16:38 crc kubenswrapper[4945]: I1206 09:16:38.803057 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:16:38 crc kubenswrapper[4945]: I1206 09:16:38.803741 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:16:42 crc kubenswrapper[4945]: I1206 09:16:42.771663 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 09:16:42 crc kubenswrapper[4945]: I1206 09:16:42.772233 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="a7804ea2-afec-4311-96f5-d5c95a13ed1f" containerName="nova-cell0-conductor-conductor" containerID="cri-o://9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3" gracePeriod=30 Dec 06 09:16:43 crc kubenswrapper[4945]: I1206 09:16:43.486893 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 09:16:43 crc kubenswrapper[4945]: I1206 09:16:43.487098 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="9e769c4b-cc65-4469-a585-5d7423bb3f87" containerName="nova-cell1-conductor-conductor" containerID="cri-o://187e17d0582fac38c827610ffc14e5cef0886eb9335d4686ad45f6f3b651c1b2" gracePeriod=30 Dec 06 09:16:43 crc kubenswrapper[4945]: I1206 09:16:43.609713 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 09:16:43 crc kubenswrapper[4945]: I1206 09:16:43.610232 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-log" containerID="cri-o://421e6fb693fb2a8656887c1a820052d1464dee0dd9b84fc91dc6ecf539a25ac9" gracePeriod=30 Dec 06 09:16:43 crc kubenswrapper[4945]: I1206 09:16:43.611178 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-api" containerID="cri-o://ddd55be65673d4137d1e955a25c2416a160f4212fbb22137802367313a580faa" gracePeriod=30 Dec 06 09:16:43 crc kubenswrapper[4945]: I1206 09:16:43.655218 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 09:16:43 crc kubenswrapper[4945]: I1206 09:16:43.655474 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="19af6ee2-c8a2-446b-a151-fe19310fa749" containerName="nova-scheduler-scheduler" containerID="cri-o://cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1" gracePeriod=30 Dec 06 09:16:43 crc kubenswrapper[4945]: I1206 09:16:43.682949 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 09:16:43 crc kubenswrapper[4945]: I1206 09:16:43.683237 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-log" containerID="cri-o://bedff4aff87aa6756de152dac23aeef4b9a193a18406e52672a49e628dd6fdd3" gracePeriod=30 Dec 06 09:16:43 crc kubenswrapper[4945]: I1206 09:16:43.683300 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-metadata" containerID="cri-o://d4960c3494a91d69c9f73b94d62151f7ed6516320a8f9a487843b140b1049866" gracePeriod=30 Dec 06 09:16:44 crc kubenswrapper[4945]: I1206 09:16:44.098720 4945 generic.go:334] "Generic (PLEG): container finished" podID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerID="bedff4aff87aa6756de152dac23aeef4b9a193a18406e52672a49e628dd6fdd3" exitCode=143 Dec 06 09:16:44 crc kubenswrapper[4945]: I1206 09:16:44.098779 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7","Type":"ContainerDied","Data":"bedff4aff87aa6756de152dac23aeef4b9a193a18406e52672a49e628dd6fdd3"} Dec 06 09:16:44 crc kubenswrapper[4945]: I1206 09:16:44.101134 4945 generic.go:334] "Generic (PLEG): container finished" podID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerID="421e6fb693fb2a8656887c1a820052d1464dee0dd9b84fc91dc6ecf539a25ac9" exitCode=143 Dec 06 09:16:44 crc kubenswrapper[4945]: I1206 09:16:44.101170 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1907aea-7d79-4ed6-a35f-a8a501e4c03f","Type":"ContainerDied","Data":"421e6fb693fb2a8656887c1a820052d1464dee0dd9b84fc91dc6ecf539a25ac9"} Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.113097 4945 generic.go:334] "Generic (PLEG): container finished" podID="9e769c4b-cc65-4469-a585-5d7423bb3f87" containerID="187e17d0582fac38c827610ffc14e5cef0886eb9335d4686ad45f6f3b651c1b2" exitCode=0 Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.113463 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9e769c4b-cc65-4469-a585-5d7423bb3f87","Type":"ContainerDied","Data":"187e17d0582fac38c827610ffc14e5cef0886eb9335d4686ad45f6f3b651c1b2"} Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.368916 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.483025 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-combined-ca-bundle\") pod \"9e769c4b-cc65-4469-a585-5d7423bb3f87\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.483228 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tblxm\" (UniqueName: \"kubernetes.io/projected/9e769c4b-cc65-4469-a585-5d7423bb3f87-kube-api-access-tblxm\") pod \"9e769c4b-cc65-4469-a585-5d7423bb3f87\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.483314 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-config-data\") pod \"9e769c4b-cc65-4469-a585-5d7423bb3f87\" (UID: \"9e769c4b-cc65-4469-a585-5d7423bb3f87\") " Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.488589 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e769c4b-cc65-4469-a585-5d7423bb3f87-kube-api-access-tblxm" (OuterVolumeSpecName: "kube-api-access-tblxm") pod "9e769c4b-cc65-4469-a585-5d7423bb3f87" (UID: "9e769c4b-cc65-4469-a585-5d7423bb3f87"). InnerVolumeSpecName "kube-api-access-tblxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.510103 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-config-data" (OuterVolumeSpecName: "config-data") pod "9e769c4b-cc65-4469-a585-5d7423bb3f87" (UID: "9e769c4b-cc65-4469-a585-5d7423bb3f87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.511865 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e769c4b-cc65-4469-a585-5d7423bb3f87" (UID: "9e769c4b-cc65-4469-a585-5d7423bb3f87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.585910 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.585946 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tblxm\" (UniqueName: \"kubernetes.io/projected/9e769c4b-cc65-4469-a585-5d7423bb3f87-kube-api-access-tblxm\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:45 crc kubenswrapper[4945]: I1206 09:16:45.585959 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e769c4b-cc65-4469-a585-5d7423bb3f87-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.124087 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9e769c4b-cc65-4469-a585-5d7423bb3f87","Type":"ContainerDied","Data":"a1fc1d704317c11bcb666d1c69af66afd7bd8e437868404c30a5f2144839de0d"} Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.124142 4945 scope.go:117] "RemoveContainer" containerID="187e17d0582fac38c827610ffc14e5cef0886eb9335d4686ad45f6f3b651c1b2" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.124200 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.164405 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.177473 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.193447 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 09:16:46 crc kubenswrapper[4945]: E1206 09:16:46.193981 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f979eb-bec9-48f7-b87f-f9618e999665" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.194005 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f979eb-bec9-48f7-b87f-f9618e999665" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 06 09:16:46 crc kubenswrapper[4945]: E1206 09:16:46.194044 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b685769-099a-4def-a772-82adbde350b1" containerName="registry-server" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.194053 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b685769-099a-4def-a772-82adbde350b1" containerName="registry-server" Dec 06 09:16:46 crc kubenswrapper[4945]: E1206 09:16:46.194075 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b685769-099a-4def-a772-82adbde350b1" containerName="extract-content" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.194083 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b685769-099a-4def-a772-82adbde350b1" containerName="extract-content" Dec 06 09:16:46 crc kubenswrapper[4945]: E1206 09:16:46.194098 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e769c4b-cc65-4469-a585-5d7423bb3f87" containerName="nova-cell1-conductor-conductor" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.194105 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e769c4b-cc65-4469-a585-5d7423bb3f87" containerName="nova-cell1-conductor-conductor" Dec 06 09:16:46 crc kubenswrapper[4945]: E1206 09:16:46.194121 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b685769-099a-4def-a772-82adbde350b1" containerName="extract-utilities" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.194130 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b685769-099a-4def-a772-82adbde350b1" containerName="extract-utilities" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.194395 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e769c4b-cc65-4469-a585-5d7423bb3f87" containerName="nova-cell1-conductor-conductor" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.194444 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="94f979eb-bec9-48f7-b87f-f9618e999665" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.194457 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b685769-099a-4def-a772-82adbde350b1" containerName="registry-server" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.195326 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.200107 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.221663 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.301976 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2d829-66ae-404a-9802-f7c696bf8d7d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9ef2d829-66ae-404a-9802-f7c696bf8d7d\") " pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.302021 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef2d829-66ae-404a-9802-f7c696bf8d7d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9ef2d829-66ae-404a-9802-f7c696bf8d7d\") " pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.302097 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjpqg\" (UniqueName: \"kubernetes.io/projected/9ef2d829-66ae-404a-9802-f7c696bf8d7d-kube-api-access-vjpqg\") pod \"nova-cell1-conductor-0\" (UID: \"9ef2d829-66ae-404a-9802-f7c696bf8d7d\") " pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.404297 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2d829-66ae-404a-9802-f7c696bf8d7d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9ef2d829-66ae-404a-9802-f7c696bf8d7d\") " pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.404354 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef2d829-66ae-404a-9802-f7c696bf8d7d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9ef2d829-66ae-404a-9802-f7c696bf8d7d\") " pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.404428 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjpqg\" (UniqueName: \"kubernetes.io/projected/9ef2d829-66ae-404a-9802-f7c696bf8d7d-kube-api-access-vjpqg\") pod \"nova-cell1-conductor-0\" (UID: \"9ef2d829-66ae-404a-9802-f7c696bf8d7d\") " pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.408905 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2d829-66ae-404a-9802-f7c696bf8d7d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9ef2d829-66ae-404a-9802-f7c696bf8d7d\") " pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.423994 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef2d829-66ae-404a-9802-f7c696bf8d7d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9ef2d829-66ae-404a-9802-f7c696bf8d7d\") " pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.429104 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjpqg\" (UniqueName: \"kubernetes.io/projected/9ef2d829-66ae-404a-9802-f7c696bf8d7d-kube-api-access-vjpqg\") pod \"nova-cell1-conductor-0\" (UID: \"9ef2d829-66ae-404a-9802-f7c696bf8d7d\") " pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.529234 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:46 crc kubenswrapper[4945]: I1206 09:16:46.966765 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e769c4b-cc65-4469-a585-5d7423bb3f87" path="/var/lib/kubelet/pods/9e769c4b-cc65-4469-a585-5d7423bb3f87/volumes" Dec 06 09:16:47 crc kubenswrapper[4945]: I1206 09:16:47.039879 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.80:8774/\": read tcp 10.217.0.2:59562->10.217.1.80:8774: read: connection reset by peer" Dec 06 09:16:47 crc kubenswrapper[4945]: I1206 09:16:47.040223 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.80:8774/\": read tcp 10.217.0.2:59560->10.217.1.80:8774: read: connection reset by peer" Dec 06 09:16:47 crc kubenswrapper[4945]: I1206 09:16:47.050069 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 09:16:47 crc kubenswrapper[4945]: I1206 09:16:47.138791 4945 generic.go:334] "Generic (PLEG): container finished" podID="19af6ee2-c8a2-446b-a151-fe19310fa749" containerID="cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1" exitCode=0 Dec 06 09:16:47 crc kubenswrapper[4945]: I1206 09:16:47.138891 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19af6ee2-c8a2-446b-a151-fe19310fa749","Type":"ContainerDied","Data":"cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1"} Dec 06 09:16:47 crc kubenswrapper[4945]: I1206 09:16:47.141533 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9ef2d829-66ae-404a-9802-f7c696bf8d7d","Type":"ContainerStarted","Data":"dd001cc7ac7aa6eb809e08e12659aee33ef18fde7849f387125224d19c45b793"} Dec 06 09:16:47 crc kubenswrapper[4945]: E1206 09:16:47.751609 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3 is running failed: container process not found" containerID="9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 09:16:47 crc kubenswrapper[4945]: E1206 09:16:47.753183 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3 is running failed: container process not found" containerID="9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 09:16:47 crc kubenswrapper[4945]: E1206 09:16:47.754778 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3 is running failed: container process not found" containerID="9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 06 09:16:47 crc kubenswrapper[4945]: E1206 09:16:47.754813 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="a7804ea2-afec-4311-96f5-d5c95a13ed1f" containerName="nova-cell0-conductor-conductor" Dec 06 09:16:47 crc kubenswrapper[4945]: I1206 09:16:47.792035 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": read tcp 10.217.0.2:35578->10.217.1.81:8775: read: connection reset by peer" Dec 06 09:16:47 crc kubenswrapper[4945]: I1206 09:16:47.792430 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": read tcp 10.217.0.2:35566->10.217.1.81:8775: read: connection reset by peer" Dec 06 09:16:47 crc kubenswrapper[4945]: E1206 09:16:47.843690 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1 is running failed: container process not found" containerID="cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 09:16:47 crc kubenswrapper[4945]: E1206 09:16:47.844064 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1 is running failed: container process not found" containerID="cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 09:16:47 crc kubenswrapper[4945]: E1206 09:16:47.845884 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1 is running failed: container process not found" containerID="cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 09:16:47 crc kubenswrapper[4945]: E1206 09:16:47.845926 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="19af6ee2-c8a2-446b-a151-fe19310fa749" containerName="nova-scheduler-scheduler" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.172002 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.176703 4945 generic.go:334] "Generic (PLEG): container finished" podID="a7804ea2-afec-4311-96f5-d5c95a13ed1f" containerID="9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3" exitCode=0 Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.176792 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a7804ea2-afec-4311-96f5-d5c95a13ed1f","Type":"ContainerDied","Data":"9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3"} Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.179841 4945 generic.go:334] "Generic (PLEG): container finished" podID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerID="d4960c3494a91d69c9f73b94d62151f7ed6516320a8f9a487843b140b1049866" exitCode=0 Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.179891 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7","Type":"ContainerDied","Data":"d4960c3494a91d69c9f73b94d62151f7ed6516320a8f9a487843b140b1049866"} Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.187113 4945 generic.go:334] "Generic (PLEG): container finished" podID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerID="ddd55be65673d4137d1e955a25c2416a160f4212fbb22137802367313a580faa" exitCode=0 Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.187155 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1907aea-7d79-4ed6-a35f-a8a501e4c03f","Type":"ContainerDied","Data":"ddd55be65673d4137d1e955a25c2416a160f4212fbb22137802367313a580faa"} Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.193234 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19af6ee2-c8a2-446b-a151-fe19310fa749","Type":"ContainerDied","Data":"6f1f0af4f152ac1e76650df0677c614e8cfb732fcb1154f0da997737f892390b"} Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.193308 4945 scope.go:117] "RemoveContainer" containerID="cfc738e128d7e47d34f04b2260557df5257a9bfa98eb35857d21e3ea931909a1" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.194399 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.263638 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg9bl\" (UniqueName: \"kubernetes.io/projected/19af6ee2-c8a2-446b-a151-fe19310fa749-kube-api-access-rg9bl\") pod \"19af6ee2-c8a2-446b-a151-fe19310fa749\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.263922 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-config-data\") pod \"19af6ee2-c8a2-446b-a151-fe19310fa749\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.264102 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-combined-ca-bundle\") pod \"19af6ee2-c8a2-446b-a151-fe19310fa749\" (UID: \"19af6ee2-c8a2-446b-a151-fe19310fa749\") " Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.277716 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19af6ee2-c8a2-446b-a151-fe19310fa749-kube-api-access-rg9bl" (OuterVolumeSpecName: "kube-api-access-rg9bl") pod "19af6ee2-c8a2-446b-a151-fe19310fa749" (UID: "19af6ee2-c8a2-446b-a151-fe19310fa749"). InnerVolumeSpecName "kube-api-access-rg9bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.309147 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm"] Dec 06 09:16:48 crc kubenswrapper[4945]: E1206 09:16:48.309626 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19af6ee2-c8a2-446b-a151-fe19310fa749" containerName="nova-scheduler-scheduler" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.309645 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="19af6ee2-c8a2-446b-a151-fe19310fa749" containerName="nova-scheduler-scheduler" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.309840 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="19af6ee2-c8a2-446b-a151-fe19310fa749" containerName="nova-scheduler-scheduler" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.310595 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.313686 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-kfm42" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.314240 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.314501 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.314745 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.314988 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.316063 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.316192 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.330059 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm"] Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.330155 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-config-data" (OuterVolumeSpecName: "config-data") pod "19af6ee2-c8a2-446b-a151-fe19310fa749" (UID: "19af6ee2-c8a2-446b-a151-fe19310fa749"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.334887 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19af6ee2-c8a2-446b-a151-fe19310fa749" (UID: "19af6ee2-c8a2-446b-a151-fe19310fa749"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.366754 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg9bl\" (UniqueName: \"kubernetes.io/projected/19af6ee2-c8a2-446b-a151-fe19310fa749-kube-api-access-rg9bl\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.366785 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.366794 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19af6ee2-c8a2-446b-a151-fe19310fa749-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468312 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468375 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468405 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468422 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468458 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468528 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468715 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468798 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468870 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnclk\" (UniqueName: \"kubernetes.io/projected/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-kube-api-access-lnclk\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468930 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.468982 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.539088 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.551926 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572062 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572159 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572192 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572216 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572265 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572313 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572344 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572379 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572413 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnclk\" (UniqueName: \"kubernetes.io/projected/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-kube-api-access-lnclk\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572455 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.572485 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.578450 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.579310 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.579365 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.579855 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.580560 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.581312 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.581461 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.583557 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.583768 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.586517 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.588443 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.589532 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.598050 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.609502 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.616483 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnclk\" (UniqueName: \"kubernetes.io/projected/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-kube-api-access-lnclk\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.655039 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.675155 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e224dc5-9c91-4441-8634-3e43c8cf89e0-config-data\") pod \"nova-scheduler-0\" (UID: \"5e224dc5-9c91-4441-8634-3e43c8cf89e0\") " pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.675217 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e224dc5-9c91-4441-8634-3e43c8cf89e0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5e224dc5-9c91-4441-8634-3e43c8cf89e0\") " pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.675351 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4gtq\" (UniqueName: \"kubernetes.io/projected/5e224dc5-9c91-4441-8634-3e43c8cf89e0-kube-api-access-z4gtq\") pod \"nova-scheduler-0\" (UID: \"5e224dc5-9c91-4441-8634-3e43c8cf89e0\") " pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.777345 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4gtq\" (UniqueName: \"kubernetes.io/projected/5e224dc5-9c91-4441-8634-3e43c8cf89e0-kube-api-access-z4gtq\") pod \"nova-scheduler-0\" (UID: \"5e224dc5-9c91-4441-8634-3e43c8cf89e0\") " pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.777535 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e224dc5-9c91-4441-8634-3e43c8cf89e0-config-data\") pod \"nova-scheduler-0\" (UID: \"5e224dc5-9c91-4441-8634-3e43c8cf89e0\") " pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.777588 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e224dc5-9c91-4441-8634-3e43c8cf89e0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5e224dc5-9c91-4441-8634-3e43c8cf89e0\") " pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.786257 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e224dc5-9c91-4441-8634-3e43c8cf89e0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5e224dc5-9c91-4441-8634-3e43c8cf89e0\") " pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.786800 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e224dc5-9c91-4441-8634-3e43c8cf89e0-config-data\") pod \"nova-scheduler-0\" (UID: \"5e224dc5-9c91-4441-8634-3e43c8cf89e0\") " pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.809165 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4gtq\" (UniqueName: \"kubernetes.io/projected/5e224dc5-9c91-4441-8634-3e43c8cf89e0-kube-api-access-z4gtq\") pod \"nova-scheduler-0\" (UID: \"5e224dc5-9c91-4441-8634-3e43c8cf89e0\") " pod="openstack/nova-scheduler-0" Dec 06 09:16:48 crc kubenswrapper[4945]: I1206 09:16:48.975358 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19af6ee2-c8a2-446b-a151-fe19310fa749" path="/var/lib/kubelet/pods/19af6ee2-c8a2-446b-a151-fe19310fa749/volumes" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.025707 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.039324 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.185258 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-combined-ca-bundle\") pod \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.185654 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-config-data\") pod \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.185866 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvg7z\" (UniqueName: \"kubernetes.io/projected/a7804ea2-afec-4311-96f5-d5c95a13ed1f-kube-api-access-vvg7z\") pod \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\" (UID: \"a7804ea2-afec-4311-96f5-d5c95a13ed1f\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.192759 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7804ea2-afec-4311-96f5-d5c95a13ed1f-kube-api-access-vvg7z" (OuterVolumeSpecName: "kube-api-access-vvg7z") pod "a7804ea2-afec-4311-96f5-d5c95a13ed1f" (UID: "a7804ea2-afec-4311-96f5-d5c95a13ed1f"). InnerVolumeSpecName "kube-api-access-vvg7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.246721 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7804ea2-afec-4311-96f5-d5c95a13ed1f" (UID: "a7804ea2-afec-4311-96f5-d5c95a13ed1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.253430 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-config-data" (OuterVolumeSpecName: "config-data") pod "a7804ea2-afec-4311-96f5-d5c95a13ed1f" (UID: "a7804ea2-afec-4311-96f5-d5c95a13ed1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.288325 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvg7z\" (UniqueName: \"kubernetes.io/projected/a7804ea2-afec-4311-96f5-d5c95a13ed1f-kube-api-access-vvg7z\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.288367 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.288379 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7804ea2-afec-4311-96f5-d5c95a13ed1f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.293620 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9ef2d829-66ae-404a-9802-f7c696bf8d7d","Type":"ContainerStarted","Data":"d33acd0a2a25dfe6faef46dd51c3e176c02eeb91f6bae2cb47b6a14e90c5f04b"} Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.293978 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.310950 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a7804ea2-afec-4311-96f5-d5c95a13ed1f","Type":"ContainerDied","Data":"447a4fc05de903eaec22613235f611cc48e2a5c92a232cd5cfd6c75e9e620882"} Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.311015 4945 scope.go:117] "RemoveContainer" containerID="9d25eba69cd2785815f8e12f6ffcd8f965537737fe9cef38144483501c1e5ac3" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.311173 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.358259 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.358237095 podStartE2EDuration="3.358237095s" podCreationTimestamp="2025-12-06 09:16:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 09:16:49.31615682 +0000 UTC m=+8622.771017864" watchObservedRunningTime="2025-12-06 09:16:49.358237095 +0000 UTC m=+8622.813098139" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.368751 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.392778 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-combined-ca-bundle\") pod \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.392846 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlqdd\" (UniqueName: \"kubernetes.io/projected/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-kube-api-access-xlqdd\") pod \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.392953 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-config-data\") pod \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.393036 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-logs\") pod \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\" (UID: \"f1907aea-7d79-4ed6-a35f-a8a501e4c03f\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.394739 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-logs" (OuterVolumeSpecName: "logs") pod "f1907aea-7d79-4ed6-a35f-a8a501e4c03f" (UID: "f1907aea-7d79-4ed6-a35f-a8a501e4c03f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.401371 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-kube-api-access-xlqdd" (OuterVolumeSpecName: "kube-api-access-xlqdd") pod "f1907aea-7d79-4ed6-a35f-a8a501e4c03f" (UID: "f1907aea-7d79-4ed6-a35f-a8a501e4c03f"). InnerVolumeSpecName "kube-api-access-xlqdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.455514 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1907aea-7d79-4ed6-a35f-a8a501e4c03f" (UID: "f1907aea-7d79-4ed6-a35f-a8a501e4c03f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.472789 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.484678 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.488095 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-config-data" (OuterVolumeSpecName: "config-data") pod "f1907aea-7d79-4ed6-a35f-a8a501e4c03f" (UID: "f1907aea-7d79-4ed6-a35f-a8a501e4c03f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.495814 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.496100 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.496126 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlqdd\" (UniqueName: \"kubernetes.io/projected/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-kube-api-access-xlqdd\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.496138 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.496151 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1907aea-7d79-4ed6-a35f-a8a501e4c03f-logs\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: E1206 09:16:49.496260 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-api" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.496294 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-api" Dec 06 09:16:49 crc kubenswrapper[4945]: E1206 09:16:49.496314 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-log" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.496322 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-log" Dec 06 09:16:49 crc kubenswrapper[4945]: E1206 09:16:49.496360 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7804ea2-afec-4311-96f5-d5c95a13ed1f" containerName="nova-cell0-conductor-conductor" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.496369 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7804ea2-afec-4311-96f5-d5c95a13ed1f" containerName="nova-cell0-conductor-conductor" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.497138 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7804ea2-afec-4311-96f5-d5c95a13ed1f" containerName="nova-cell0-conductor-conductor" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.497160 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-log" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.497173 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" containerName="nova-api-api" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.498491 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.500564 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.526470 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.595000 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.600805 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e477bf3-d8bf-47ae-9b81-080d365397ac-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0e477bf3-d8bf-47ae-9b81-080d365397ac\") " pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.600965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e477bf3-d8bf-47ae-9b81-080d365397ac-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0e477bf3-d8bf-47ae-9b81-080d365397ac\") " pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.600997 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp6dx\" (UniqueName: \"kubernetes.io/projected/0e477bf3-d8bf-47ae-9b81-080d365397ac-kube-api-access-mp6dx\") pod \"nova-cell0-conductor-0\" (UID: \"0e477bf3-d8bf-47ae-9b81-080d365397ac\") " pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.699016 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm"] Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.702859 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-combined-ca-bundle\") pod \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.702945 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-config-data\") pod \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.703061 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-logs\") pod \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.703254 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz796\" (UniqueName: \"kubernetes.io/projected/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-kube-api-access-wz796\") pod \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\" (UID: \"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7\") " Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.703625 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e477bf3-d8bf-47ae-9b81-080d365397ac-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0e477bf3-d8bf-47ae-9b81-080d365397ac\") " pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.703665 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp6dx\" (UniqueName: \"kubernetes.io/projected/0e477bf3-d8bf-47ae-9b81-080d365397ac-kube-api-access-mp6dx\") pod \"nova-cell0-conductor-0\" (UID: \"0e477bf3-d8bf-47ae-9b81-080d365397ac\") " pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.703870 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e477bf3-d8bf-47ae-9b81-080d365397ac-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0e477bf3-d8bf-47ae-9b81-080d365397ac\") " pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.704470 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-logs" (OuterVolumeSpecName: "logs") pod "d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" (UID: "d66d0ddf-dfbc-47aa-9e57-a8405e8609b7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: W1206 09:16:49.719548 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5c4bb2e_c6c7_49f4_8c53_32e4dba43b4c.slice/crio-b6204177dcef93da4f082f1e56c488dd1bc4fbefd16b6252a9abf4b8f8ef7bba WatchSource:0}: Error finding container b6204177dcef93da4f082f1e56c488dd1bc4fbefd16b6252a9abf4b8f8ef7bba: Status 404 returned error can't find the container with id b6204177dcef93da4f082f1e56c488dd1bc4fbefd16b6252a9abf4b8f8ef7bba Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.726024 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e477bf3-d8bf-47ae-9b81-080d365397ac-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0e477bf3-d8bf-47ae-9b81-080d365397ac\") " pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.728960 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-kube-api-access-wz796" (OuterVolumeSpecName: "kube-api-access-wz796") pod "d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" (UID: "d66d0ddf-dfbc-47aa-9e57-a8405e8609b7"). InnerVolumeSpecName "kube-api-access-wz796". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.729900 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e477bf3-d8bf-47ae-9b81-080d365397ac-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0e477bf3-d8bf-47ae-9b81-080d365397ac\") " pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.743003 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp6dx\" (UniqueName: \"kubernetes.io/projected/0e477bf3-d8bf-47ae-9b81-080d365397ac-kube-api-access-mp6dx\") pod \"nova-cell0-conductor-0\" (UID: \"0e477bf3-d8bf-47ae-9b81-080d365397ac\") " pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.763416 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" (UID: "d66d0ddf-dfbc-47aa-9e57-a8405e8609b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.798648 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-config-data" (OuterVolumeSpecName: "config-data") pod "d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" (UID: "d66d0ddf-dfbc-47aa-9e57-a8405e8609b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.805626 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-logs\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.805660 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz796\" (UniqueName: \"kubernetes.io/projected/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-kube-api-access-wz796\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.805674 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.805683 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.835760 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 09:16:49 crc kubenswrapper[4945]: I1206 09:16:49.921264 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.327604 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.327603 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d66d0ddf-dfbc-47aa-9e57-a8405e8609b7","Type":"ContainerDied","Data":"e4afa7894da57726fcb7929f4d927005caca2809ac36e650867630fc3e9c0591"} Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.328899 4945 scope.go:117] "RemoveContainer" containerID="d4960c3494a91d69c9f73b94d62151f7ed6516320a8f9a487843b140b1049866" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.333432 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.333492 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1907aea-7d79-4ed6-a35f-a8a501e4c03f","Type":"ContainerDied","Data":"e2ff44ace196d1fc64d39241889b03c2175ca922301a7ae6f2e6ec74d57817dc"} Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.425947 4945 scope.go:117] "RemoveContainer" containerID="bedff4aff87aa6756de152dac23aeef4b9a193a18406e52672a49e628dd6fdd3" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.447505 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" event={"ID":"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c","Type":"ContainerStarted","Data":"b6204177dcef93da4f082f1e56c488dd1bc4fbefd16b6252a9abf4b8f8ef7bba"} Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.474121 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5e224dc5-9c91-4441-8634-3e43c8cf89e0","Type":"ContainerStarted","Data":"8fbb76a3c6188d7d1db4edabf741889eca67f9ea100977ae47551c4feaaff4d6"} Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.474215 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5e224dc5-9c91-4441-8634-3e43c8cf89e0","Type":"ContainerStarted","Data":"da9befdb0c0e2b56b6c731e8251b2d2583a600e43622c68ac6980d39830cd8ec"} Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.498634 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.516790 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.5167657 podStartE2EDuration="2.5167657s" podCreationTimestamp="2025-12-06 09:16:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 09:16:50.489340607 +0000 UTC m=+8623.944201661" watchObservedRunningTime="2025-12-06 09:16:50.5167657 +0000 UTC m=+8623.971626744" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.705610 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.711047 4945 scope.go:117] "RemoveContainer" containerID="ddd55be65673d4137d1e955a25c2416a160f4212fbb22137802367313a580faa" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.730840 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.744715 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.788547 4945 scope.go:117] "RemoveContainer" containerID="421e6fb693fb2a8656887c1a820052d1464dee0dd9b84fc91dc6ecf539a25ac9" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.790164 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 09:16:50 crc kubenswrapper[4945]: E1206 09:16:50.790628 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-metadata" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.790647 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-metadata" Dec 06 09:16:50 crc kubenswrapper[4945]: E1206 09:16:50.790677 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-log" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.790694 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-log" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.790940 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-log" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.790981 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" containerName="nova-metadata-metadata" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.792906 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.805670 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.809501 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.823803 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.834509 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pllb\" (UniqueName: \"kubernetes.io/projected/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-kube-api-access-5pllb\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.834742 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.834856 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-config-data\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.838911 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-logs\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.854123 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.856072 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.858888 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.873227 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.940800 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwwsk\" (UniqueName: \"kubernetes.io/projected/0ec077cb-3c84-4617-bd63-6333112b514e-kube-api-access-kwwsk\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.940921 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ec077cb-3c84-4617-bd63-6333112b514e-logs\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.941008 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-logs\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.941089 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec077cb-3c84-4617-bd63-6333112b514e-config-data\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.941251 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pllb\" (UniqueName: \"kubernetes.io/projected/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-kube-api-access-5pllb\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.941297 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.941332 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec077cb-3c84-4617-bd63-6333112b514e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.942007 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-config-data\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.946674 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-logs\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.950615 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-config-data\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.974571 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pllb\" (UniqueName: \"kubernetes.io/projected/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-kube-api-access-5pllb\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.974823 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18b0940d-3e8f-4d9d-ab7c-530e5da63e4d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d\") " pod="openstack/nova-metadata-0" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.987502 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7804ea2-afec-4311-96f5-d5c95a13ed1f" path="/var/lib/kubelet/pods/a7804ea2-afec-4311-96f5-d5c95a13ed1f/volumes" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.995784 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d66d0ddf-dfbc-47aa-9e57-a8405e8609b7" path="/var/lib/kubelet/pods/d66d0ddf-dfbc-47aa-9e57-a8405e8609b7/volumes" Dec 06 09:16:50 crc kubenswrapper[4945]: I1206 09:16:50.996911 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1907aea-7d79-4ed6-a35f-a8a501e4c03f" path="/var/lib/kubelet/pods/f1907aea-7d79-4ed6-a35f-a8a501e4c03f/volumes" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.044160 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ec077cb-3c84-4617-bd63-6333112b514e-logs\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.044365 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec077cb-3c84-4617-bd63-6333112b514e-config-data\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.044643 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec077cb-3c84-4617-bd63-6333112b514e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.044835 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwwsk\" (UniqueName: \"kubernetes.io/projected/0ec077cb-3c84-4617-bd63-6333112b514e-kube-api-access-kwwsk\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.046323 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ec077cb-3c84-4617-bd63-6333112b514e-logs\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.052455 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec077cb-3c84-4617-bd63-6333112b514e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.052644 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec077cb-3c84-4617-bd63-6333112b514e-config-data\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.064208 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwwsk\" (UniqueName: \"kubernetes.io/projected/0ec077cb-3c84-4617-bd63-6333112b514e-kube-api-access-kwwsk\") pod \"nova-api-0\" (UID: \"0ec077cb-3c84-4617-bd63-6333112b514e\") " pod="openstack/nova-api-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.134237 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.182247 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.491011 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0e477bf3-d8bf-47ae-9b81-080d365397ac","Type":"ContainerStarted","Data":"1e875f674c44755587ec7ba003b89feae2a2df8a23b79573f02e69a10651e91a"} Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.491394 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0e477bf3-d8bf-47ae-9b81-080d365397ac","Type":"ContainerStarted","Data":"9d184ce6e3eebe5b20ca4e8194486e478ecc68127c3da08f42beacb2e945bb99"} Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.491458 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.505908 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" event={"ID":"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c","Type":"ContainerStarted","Data":"ff9975509b460b5d619b49f02bbfc3db5632e6c91067395ea244721b18330067"} Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.533542 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.533522306 podStartE2EDuration="2.533522306s" podCreationTimestamp="2025-12-06 09:16:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 09:16:51.529470017 +0000 UTC m=+8624.984331061" watchObservedRunningTime="2025-12-06 09:16:51.533522306 +0000 UTC m=+8624.988383350" Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.547925 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.576761 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" podStartSLOduration=3.077583 podStartE2EDuration="3.576737851s" podCreationTimestamp="2025-12-06 09:16:48 +0000 UTC" firstStartedPulling="2025-12-06 09:16:49.72809905 +0000 UTC m=+8623.182960094" lastFinishedPulling="2025-12-06 09:16:50.227253901 +0000 UTC m=+8623.682114945" observedRunningTime="2025-12-06 09:16:51.574792499 +0000 UTC m=+8625.029653553" watchObservedRunningTime="2025-12-06 09:16:51.576737851 +0000 UTC m=+8625.031598895" Dec 06 09:16:51 crc kubenswrapper[4945]: W1206 09:16:51.670688 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18b0940d_3e8f_4d9d_ab7c_530e5da63e4d.slice/crio-2673d215d23f078212911b5dea06766ba4c3f31006f6c010a773a05c6b3ca892 WatchSource:0}: Error finding container 2673d215d23f078212911b5dea06766ba4c3f31006f6c010a773a05c6b3ca892: Status 404 returned error can't find the container with id 2673d215d23f078212911b5dea06766ba4c3f31006f6c010a773a05c6b3ca892 Dec 06 09:16:51 crc kubenswrapper[4945]: I1206 09:16:51.828785 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 09:16:52 crc kubenswrapper[4945]: I1206 09:16:52.534951 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d","Type":"ContainerStarted","Data":"966b73524e8c6e33085309d91ee99f3516ca4f3177555c8a6d564ad39dcd5472"} Dec 06 09:16:52 crc kubenswrapper[4945]: I1206 09:16:52.535292 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d","Type":"ContainerStarted","Data":"b882f2b12e61b6bdc6d8c4415df3010abe4e77a937684d54ba8ff29ccfe37413"} Dec 06 09:16:52 crc kubenswrapper[4945]: I1206 09:16:52.535309 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18b0940d-3e8f-4d9d-ab7c-530e5da63e4d","Type":"ContainerStarted","Data":"2673d215d23f078212911b5dea06766ba4c3f31006f6c010a773a05c6b3ca892"} Dec 06 09:16:52 crc kubenswrapper[4945]: I1206 09:16:52.537514 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ec077cb-3c84-4617-bd63-6333112b514e","Type":"ContainerStarted","Data":"6040541e34b4e5c1ad12860aa294b5073c36a665778fff09eaa8efadc0be665f"} Dec 06 09:16:52 crc kubenswrapper[4945]: I1206 09:16:52.537586 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ec077cb-3c84-4617-bd63-6333112b514e","Type":"ContainerStarted","Data":"ecfce5a992743e715e75d246dafad59685792916a87ef9e6e7d6275b7a33c04f"} Dec 06 09:16:52 crc kubenswrapper[4945]: I1206 09:16:52.537602 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ec077cb-3c84-4617-bd63-6333112b514e","Type":"ContainerStarted","Data":"c3e2f83fc68b80b5a09ae246d66f579cc4d6efdbf64d720dd3cca860289605f0"} Dec 06 09:16:52 crc kubenswrapper[4945]: I1206 09:16:52.568928 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.56890737 podStartE2EDuration="2.56890737s" podCreationTimestamp="2025-12-06 09:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 09:16:52.555725128 +0000 UTC m=+8626.010586182" watchObservedRunningTime="2025-12-06 09:16:52.56890737 +0000 UTC m=+8626.023768404" Dec 06 09:16:54 crc kubenswrapper[4945]: I1206 09:16:54.026422 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 09:16:56 crc kubenswrapper[4945]: I1206 09:16:56.134332 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 09:16:56 crc kubenswrapper[4945]: I1206 09:16:56.134466 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 09:16:56 crc kubenswrapper[4945]: I1206 09:16:56.560231 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 06 09:16:56 crc kubenswrapper[4945]: I1206 09:16:56.581052 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=6.581032715 podStartE2EDuration="6.581032715s" podCreationTimestamp="2025-12-06 09:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 09:16:52.591803322 +0000 UTC m=+8626.046664366" watchObservedRunningTime="2025-12-06 09:16:56.581032715 +0000 UTC m=+8630.035893779" Dec 06 09:16:59 crc kubenswrapper[4945]: I1206 09:16:59.027506 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 09:16:59 crc kubenswrapper[4945]: I1206 09:16:59.057370 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 09:16:59 crc kubenswrapper[4945]: I1206 09:16:59.637462 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 09:16:59 crc kubenswrapper[4945]: I1206 09:16:59.950348 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 06 09:17:01 crc kubenswrapper[4945]: I1206 09:17:01.134564 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 09:17:01 crc kubenswrapper[4945]: I1206 09:17:01.134610 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 09:17:01 crc kubenswrapper[4945]: I1206 09:17:01.182861 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 09:17:01 crc kubenswrapper[4945]: I1206 09:17:01.182912 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 09:17:02 crc kubenswrapper[4945]: I1206 09:17:02.216519 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="18b0940d-3e8f-4d9d-ab7c-530e5da63e4d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.175:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 09:17:02 crc kubenswrapper[4945]: I1206 09:17:02.216463 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="18b0940d-3e8f-4d9d-ab7c-530e5da63e4d" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.175:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 09:17:02 crc kubenswrapper[4945]: I1206 09:17:02.257520 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0ec077cb-3c84-4617-bd63-6333112b514e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.176:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 09:17:02 crc kubenswrapper[4945]: I1206 09:17:02.299492 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0ec077cb-3c84-4617-bd63-6333112b514e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.176:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 09:17:08 crc kubenswrapper[4945]: I1206 09:17:08.795961 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:17:08 crc kubenswrapper[4945]: I1206 09:17:08.796598 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:17:11 crc kubenswrapper[4945]: I1206 09:17:11.143806 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 09:17:11 crc kubenswrapper[4945]: I1206 09:17:11.145852 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 09:17:11 crc kubenswrapper[4945]: I1206 09:17:11.146074 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 09:17:11 crc kubenswrapper[4945]: I1206 09:17:11.193450 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 09:17:11 crc kubenswrapper[4945]: I1206 09:17:11.193521 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 09:17:11 crc kubenswrapper[4945]: I1206 09:17:11.194916 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 09:17:11 crc kubenswrapper[4945]: I1206 09:17:11.194957 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 09:17:11 crc kubenswrapper[4945]: I1206 09:17:11.204724 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 09:17:11 crc kubenswrapper[4945]: I1206 09:17:11.206094 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 09:17:11 crc kubenswrapper[4945]: I1206 09:17:11.738195 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 09:17:38 crc kubenswrapper[4945]: I1206 09:17:38.796205 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:17:38 crc kubenswrapper[4945]: I1206 09:17:38.796714 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:17:38 crc kubenswrapper[4945]: I1206 09:17:38.796761 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 09:17:38 crc kubenswrapper[4945]: I1206 09:17:38.797359 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6971dd6d12e521980a0aacdcb6dd8e0a16a4b48aeec4b9057f6e6d63ff4f22c6"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 09:17:38 crc kubenswrapper[4945]: I1206 09:17:38.797423 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://6971dd6d12e521980a0aacdcb6dd8e0a16a4b48aeec4b9057f6e6d63ff4f22c6" gracePeriod=600 Dec 06 09:17:39 crc kubenswrapper[4945]: I1206 09:17:39.003303 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="6971dd6d12e521980a0aacdcb6dd8e0a16a4b48aeec4b9057f6e6d63ff4f22c6" exitCode=0 Dec 06 09:17:39 crc kubenswrapper[4945]: I1206 09:17:39.003386 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"6971dd6d12e521980a0aacdcb6dd8e0a16a4b48aeec4b9057f6e6d63ff4f22c6"} Dec 06 09:17:39 crc kubenswrapper[4945]: I1206 09:17:39.003655 4945 scope.go:117] "RemoveContainer" containerID="b247f9ab734715d55e5eb950fd1a6a35706c500f1346af1cfc62fcf11cb8f99f" Dec 06 09:17:40 crc kubenswrapper[4945]: I1206 09:17:40.016824 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014"} Dec 06 09:20:08 crc kubenswrapper[4945]: I1206 09:20:08.796185 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:20:08 crc kubenswrapper[4945]: I1206 09:20:08.796815 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:20:38 crc kubenswrapper[4945]: I1206 09:20:38.796098 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:20:38 crc kubenswrapper[4945]: I1206 09:20:38.796687 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:20:46 crc kubenswrapper[4945]: I1206 09:20:46.054217 4945 trace.go:236] Trace[1764153341]: "Calculate volume metrics of ovndbcluster-nb-etc-ovn for pod openstack/ovsdbserver-nb-0" (06-Dec-2025 09:20:43.342) (total time: 2712ms): Dec 06 09:20:46 crc kubenswrapper[4945]: Trace[1764153341]: [2.712055188s] [2.712055188s] END Dec 06 09:20:46 crc kubenswrapper[4945]: I1206 09:20:46.077226 4945 trace.go:236] Trace[1147351009]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-lslh4" (06-Dec-2025 09:20:43.364) (total time: 2712ms): Dec 06 09:20:46 crc kubenswrapper[4945]: Trace[1147351009]: [2.712593003s] [2.712593003s] END Dec 06 09:21:08 crc kubenswrapper[4945]: I1206 09:21:08.795888 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:21:08 crc kubenswrapper[4945]: I1206 09:21:08.796338 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:21:08 crc kubenswrapper[4945]: I1206 09:21:08.796375 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 09:21:08 crc kubenswrapper[4945]: I1206 09:21:08.796910 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 09:21:08 crc kubenswrapper[4945]: I1206 09:21:08.796963 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" gracePeriod=600 Dec 06 09:21:08 crc kubenswrapper[4945]: E1206 09:21:08.924112 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:21:09 crc kubenswrapper[4945]: I1206 09:21:09.526818 4945 generic.go:334] "Generic (PLEG): container finished" podID="c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" containerID="ff9975509b460b5d619b49f02bbfc3db5632e6c91067395ea244721b18330067" exitCode=0 Dec 06 09:21:09 crc kubenswrapper[4945]: I1206 09:21:09.527221 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" event={"ID":"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c","Type":"ContainerDied","Data":"ff9975509b460b5d619b49f02bbfc3db5632e6c91067395ea244721b18330067"} Dec 06 09:21:09 crc kubenswrapper[4945]: I1206 09:21:09.530173 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" exitCode=0 Dec 06 09:21:09 crc kubenswrapper[4945]: I1206 09:21:09.530226 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014"} Dec 06 09:21:09 crc kubenswrapper[4945]: I1206 09:21:09.530266 4945 scope.go:117] "RemoveContainer" containerID="6971dd6d12e521980a0aacdcb6dd8e0a16a4b48aeec4b9057f6e6d63ff4f22c6" Dec 06 09:21:09 crc kubenswrapper[4945]: I1206 09:21:09.531035 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:21:09 crc kubenswrapper[4945]: E1206 09:21:09.531455 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:21:10 crc kubenswrapper[4945]: I1206 09:21:10.984172 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.111775 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-0\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.111888 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ceph\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.111996 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ssh-key\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.112080 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-0\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.112150 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-1\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.112214 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnclk\" (UniqueName: \"kubernetes.io/projected/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-kube-api-access-lnclk\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.112271 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-inventory\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.112360 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-0\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.112414 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-1\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.112511 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-1\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.112597 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-combined-ca-bundle\") pod \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\" (UID: \"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c\") " Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.118460 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.124463 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-kube-api-access-lnclk" (OuterVolumeSpecName: "kube-api-access-lnclk") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "kube-api-access-lnclk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.124666 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ceph" (OuterVolumeSpecName: "ceph") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.143153 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.148428 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.159425 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.159475 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.159495 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.159875 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.167440 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-inventory" (OuterVolumeSpecName: "inventory") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.178415 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" (UID: "c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215200 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215243 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215256 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215270 4945 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ceph\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215313 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215326 4945 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215337 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215349 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnclk\" (UniqueName: \"kubernetes.io/projected/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-kube-api-access-lnclk\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215360 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215372 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.215384 4945 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.567550 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" event={"ID":"c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c","Type":"ContainerDied","Data":"b6204177dcef93da4f082f1e56c488dd1bc4fbefd16b6252a9abf4b8f8ef7bba"} Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.567796 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6204177dcef93da4f082f1e56c488dd1bc4fbefd16b6252a9abf4b8f8ef7bba" Dec 06 09:21:11 crc kubenswrapper[4945]: I1206 09:21:11.567645 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm" Dec 06 09:21:21 crc kubenswrapper[4945]: I1206 09:21:21.953012 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:21:21 crc kubenswrapper[4945]: E1206 09:21:21.953702 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:21:32 crc kubenswrapper[4945]: I1206 09:21:32.924994 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2l8lw"] Dec 06 09:21:32 crc kubenswrapper[4945]: E1206 09:21:32.937389 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 06 09:21:32 crc kubenswrapper[4945]: I1206 09:21:32.937672 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 06 09:21:32 crc kubenswrapper[4945]: I1206 09:21:32.937969 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 06 09:21:32 crc kubenswrapper[4945]: I1206 09:21:32.939649 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2l8lw"] Dec 06 09:21:32 crc kubenswrapper[4945]: I1206 09:21:32.939809 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.097768 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z7sd\" (UniqueName: \"kubernetes.io/projected/5b79a891-f567-4de9-be19-fae05a6722be-kube-api-access-8z7sd\") pod \"certified-operators-2l8lw\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.098458 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-catalog-content\") pod \"certified-operators-2l8lw\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.098549 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-utilities\") pod \"certified-operators-2l8lw\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.199438 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-catalog-content\") pod \"certified-operators-2l8lw\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.199490 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-utilities\") pod \"certified-operators-2l8lw\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.199639 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z7sd\" (UniqueName: \"kubernetes.io/projected/5b79a891-f567-4de9-be19-fae05a6722be-kube-api-access-8z7sd\") pod \"certified-operators-2l8lw\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.200058 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-catalog-content\") pod \"certified-operators-2l8lw\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.200126 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-utilities\") pod \"certified-operators-2l8lw\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.220629 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z7sd\" (UniqueName: \"kubernetes.io/projected/5b79a891-f567-4de9-be19-fae05a6722be-kube-api-access-8z7sd\") pod \"certified-operators-2l8lw\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.300374 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:33 crc kubenswrapper[4945]: I1206 09:21:33.864815 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2l8lw"] Dec 06 09:21:34 crc kubenswrapper[4945]: I1206 09:21:34.795082 4945 generic.go:334] "Generic (PLEG): container finished" podID="5b79a891-f567-4de9-be19-fae05a6722be" containerID="606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc" exitCode=0 Dec 06 09:21:34 crc kubenswrapper[4945]: I1206 09:21:34.795164 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l8lw" event={"ID":"5b79a891-f567-4de9-be19-fae05a6722be","Type":"ContainerDied","Data":"606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc"} Dec 06 09:21:34 crc kubenswrapper[4945]: I1206 09:21:34.795538 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l8lw" event={"ID":"5b79a891-f567-4de9-be19-fae05a6722be","Type":"ContainerStarted","Data":"adba206b15195e3affa20cdec3283ca222ece706e67e1704a4061d5bcb76aeea"} Dec 06 09:21:34 crc kubenswrapper[4945]: I1206 09:21:34.800962 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 09:21:35 crc kubenswrapper[4945]: I1206 09:21:35.807974 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l8lw" event={"ID":"5b79a891-f567-4de9-be19-fae05a6722be","Type":"ContainerStarted","Data":"238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd"} Dec 06 09:21:35 crc kubenswrapper[4945]: I1206 09:21:35.953083 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:21:35 crc kubenswrapper[4945]: E1206 09:21:35.953475 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:21:36 crc kubenswrapper[4945]: I1206 09:21:36.817604 4945 generic.go:334] "Generic (PLEG): container finished" podID="5b79a891-f567-4de9-be19-fae05a6722be" containerID="238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd" exitCode=0 Dec 06 09:21:36 crc kubenswrapper[4945]: I1206 09:21:36.817646 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l8lw" event={"ID":"5b79a891-f567-4de9-be19-fae05a6722be","Type":"ContainerDied","Data":"238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd"} Dec 06 09:21:37 crc kubenswrapper[4945]: I1206 09:21:37.829840 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l8lw" event={"ID":"5b79a891-f567-4de9-be19-fae05a6722be","Type":"ContainerStarted","Data":"54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542"} Dec 06 09:21:37 crc kubenswrapper[4945]: I1206 09:21:37.851851 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2l8lw" podStartSLOduration=3.431567844 podStartE2EDuration="5.851827922s" podCreationTimestamp="2025-12-06 09:21:32 +0000 UTC" firstStartedPulling="2025-12-06 09:21:34.80065198 +0000 UTC m=+8908.255513024" lastFinishedPulling="2025-12-06 09:21:37.220912058 +0000 UTC m=+8910.675773102" observedRunningTime="2025-12-06 09:21:37.847407973 +0000 UTC m=+8911.302269027" watchObservedRunningTime="2025-12-06 09:21:37.851827922 +0000 UTC m=+8911.306688966" Dec 06 09:21:43 crc kubenswrapper[4945]: I1206 09:21:43.301345 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:43 crc kubenswrapper[4945]: I1206 09:21:43.301943 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:43 crc kubenswrapper[4945]: I1206 09:21:43.350236 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:43 crc kubenswrapper[4945]: I1206 09:21:43.926030 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:43 crc kubenswrapper[4945]: I1206 09:21:43.978816 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2l8lw"] Dec 06 09:21:45 crc kubenswrapper[4945]: I1206 09:21:45.906982 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2l8lw" podUID="5b79a891-f567-4de9-be19-fae05a6722be" containerName="registry-server" containerID="cri-o://54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542" gracePeriod=2 Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.421831 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.575154 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z7sd\" (UniqueName: \"kubernetes.io/projected/5b79a891-f567-4de9-be19-fae05a6722be-kube-api-access-8z7sd\") pod \"5b79a891-f567-4de9-be19-fae05a6722be\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.575497 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-utilities\") pod \"5b79a891-f567-4de9-be19-fae05a6722be\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.575659 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-catalog-content\") pod \"5b79a891-f567-4de9-be19-fae05a6722be\" (UID: \"5b79a891-f567-4de9-be19-fae05a6722be\") " Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.576469 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-utilities" (OuterVolumeSpecName: "utilities") pod "5b79a891-f567-4de9-be19-fae05a6722be" (UID: "5b79a891-f567-4de9-be19-fae05a6722be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.581070 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b79a891-f567-4de9-be19-fae05a6722be-kube-api-access-8z7sd" (OuterVolumeSpecName: "kube-api-access-8z7sd") pod "5b79a891-f567-4de9-be19-fae05a6722be" (UID: "5b79a891-f567-4de9-be19-fae05a6722be"). InnerVolumeSpecName "kube-api-access-8z7sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.677902 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z7sd\" (UniqueName: \"kubernetes.io/projected/5b79a891-f567-4de9-be19-fae05a6722be-kube-api-access-8z7sd\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.678190 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.780452 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b79a891-f567-4de9-be19-fae05a6722be" (UID: "5b79a891-f567-4de9-be19-fae05a6722be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.881172 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b79a891-f567-4de9-be19-fae05a6722be-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.919060 4945 generic.go:334] "Generic (PLEG): container finished" podID="5b79a891-f567-4de9-be19-fae05a6722be" containerID="54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542" exitCode=0 Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.919100 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l8lw" event={"ID":"5b79a891-f567-4de9-be19-fae05a6722be","Type":"ContainerDied","Data":"54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542"} Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.919126 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l8lw" event={"ID":"5b79a891-f567-4de9-be19-fae05a6722be","Type":"ContainerDied","Data":"adba206b15195e3affa20cdec3283ca222ece706e67e1704a4061d5bcb76aeea"} Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.919143 4945 scope.go:117] "RemoveContainer" containerID="54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542" Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.920394 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l8lw" Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.976806 4945 scope.go:117] "RemoveContainer" containerID="238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd" Dec 06 09:21:46 crc kubenswrapper[4945]: I1206 09:21:46.993008 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2l8lw"] Dec 06 09:21:47 crc kubenswrapper[4945]: I1206 09:21:47.001924 4945 scope.go:117] "RemoveContainer" containerID="606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc" Dec 06 09:21:47 crc kubenswrapper[4945]: I1206 09:21:47.005515 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2l8lw"] Dec 06 09:21:47 crc kubenswrapper[4945]: I1206 09:21:47.047665 4945 scope.go:117] "RemoveContainer" containerID="54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542" Dec 06 09:21:47 crc kubenswrapper[4945]: E1206 09:21:47.048207 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542\": container with ID starting with 54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542 not found: ID does not exist" containerID="54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542" Dec 06 09:21:47 crc kubenswrapper[4945]: I1206 09:21:47.048315 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542"} err="failed to get container status \"54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542\": rpc error: code = NotFound desc = could not find container \"54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542\": container with ID starting with 54c59e272b0076d62c8e551b05b004c9cdd369bbead6391957d924621627a542 not found: ID does not exist" Dec 06 09:21:47 crc kubenswrapper[4945]: I1206 09:21:47.048515 4945 scope.go:117] "RemoveContainer" containerID="238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd" Dec 06 09:21:47 crc kubenswrapper[4945]: E1206 09:21:47.049055 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd\": container with ID starting with 238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd not found: ID does not exist" containerID="238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd" Dec 06 09:21:47 crc kubenswrapper[4945]: I1206 09:21:47.049140 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd"} err="failed to get container status \"238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd\": rpc error: code = NotFound desc = could not find container \"238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd\": container with ID starting with 238f4ce7269ebc66b2cd51c6773ee2fb0e2f991b93e00ffc7aae7a0963f16dfd not found: ID does not exist" Dec 06 09:21:47 crc kubenswrapper[4945]: I1206 09:21:47.049200 4945 scope.go:117] "RemoveContainer" containerID="606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc" Dec 06 09:21:47 crc kubenswrapper[4945]: E1206 09:21:47.049564 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc\": container with ID starting with 606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc not found: ID does not exist" containerID="606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc" Dec 06 09:21:47 crc kubenswrapper[4945]: I1206 09:21:47.049691 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc"} err="failed to get container status \"606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc\": rpc error: code = NotFound desc = could not find container \"606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc\": container with ID starting with 606fd3dcd6a4ba5b90970e62167b7ead28db84fdaa8de5dfc6f30384b57b53bc not found: ID does not exist" Dec 06 09:21:48 crc kubenswrapper[4945]: I1206 09:21:48.953739 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:21:48 crc kubenswrapper[4945]: E1206 09:21:48.954628 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:21:48 crc kubenswrapper[4945]: I1206 09:21:48.967092 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b79a891-f567-4de9-be19-fae05a6722be" path="/var/lib/kubelet/pods/5b79a891-f567-4de9-be19-fae05a6722be/volumes" Dec 06 09:22:00 crc kubenswrapper[4945]: I1206 09:22:00.953030 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:22:00 crc kubenswrapper[4945]: E1206 09:22:00.953947 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:22:12 crc kubenswrapper[4945]: I1206 09:22:12.953237 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:22:12 crc kubenswrapper[4945]: E1206 09:22:12.954043 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:22:24 crc kubenswrapper[4945]: I1206 09:22:24.953197 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:22:24 crc kubenswrapper[4945]: E1206 09:22:24.954060 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:22:38 crc kubenswrapper[4945]: I1206 09:22:38.953154 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:22:38 crc kubenswrapper[4945]: E1206 09:22:38.953908 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:22:50 crc kubenswrapper[4945]: I1206 09:22:50.953969 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:22:50 crc kubenswrapper[4945]: E1206 09:22:50.954760 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:23:01 crc kubenswrapper[4945]: I1206 09:23:01.955097 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:23:01 crc kubenswrapper[4945]: E1206 09:23:01.955839 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:23:15 crc kubenswrapper[4945]: I1206 09:23:15.953788 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:23:15 crc kubenswrapper[4945]: E1206 09:23:15.954517 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:23:21 crc kubenswrapper[4945]: I1206 09:23:21.222365 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 06 09:23:21 crc kubenswrapper[4945]: I1206 09:23:21.223980 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="c5917d38-82ff-4f41-ae15-61aca63e378c" containerName="adoption" containerID="cri-o://560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b" gracePeriod=30 Dec 06 09:23:27 crc kubenswrapper[4945]: I1206 09:23:27.953174 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:23:27 crc kubenswrapper[4945]: E1206 09:23:27.953961 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:23:41 crc kubenswrapper[4945]: I1206 09:23:41.953883 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:23:41 crc kubenswrapper[4945]: E1206 09:23:41.955135 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.193824 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.259758 4945 generic.go:334] "Generic (PLEG): container finished" podID="c5917d38-82ff-4f41-ae15-61aca63e378c" containerID="560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b" exitCode=137 Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.259803 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"c5917d38-82ff-4f41-ae15-61aca63e378c","Type":"ContainerDied","Data":"560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b"} Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.259811 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.259836 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"c5917d38-82ff-4f41-ae15-61aca63e378c","Type":"ContainerDied","Data":"edc0bc4fb87fa11dd58d47562df7bd5238bbb4a4c7be2a8c55895ca7f96c2f03"} Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.259856 4945 scope.go:117] "RemoveContainer" containerID="560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.280505 4945 scope.go:117] "RemoveContainer" containerID="560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b" Dec 06 09:23:52 crc kubenswrapper[4945]: E1206 09:23:52.281093 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b\": container with ID starting with 560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b not found: ID does not exist" containerID="560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.281122 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b"} err="failed to get container status \"560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b\": rpc error: code = NotFound desc = could not find container \"560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b\": container with ID starting with 560c125de728e6519e2871102378d865fe605a42e6de76039310a0b7b6ef4b2b not found: ID does not exist" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.360910 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfrlz\" (UniqueName: \"kubernetes.io/projected/c5917d38-82ff-4f41-ae15-61aca63e378c-kube-api-access-gfrlz\") pod \"c5917d38-82ff-4f41-ae15-61aca63e378c\" (UID: \"c5917d38-82ff-4f41-ae15-61aca63e378c\") " Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.371400 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\") pod \"c5917d38-82ff-4f41-ae15-61aca63e378c\" (UID: \"c5917d38-82ff-4f41-ae15-61aca63e378c\") " Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.377487 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5917d38-82ff-4f41-ae15-61aca63e378c-kube-api-access-gfrlz" (OuterVolumeSpecName: "kube-api-access-gfrlz") pod "c5917d38-82ff-4f41-ae15-61aca63e378c" (UID: "c5917d38-82ff-4f41-ae15-61aca63e378c"). InnerVolumeSpecName "kube-api-access-gfrlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.389465 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3" (OuterVolumeSpecName: "mariadb-data") pod "c5917d38-82ff-4f41-ae15-61aca63e378c" (UID: "c5917d38-82ff-4f41-ae15-61aca63e378c"). InnerVolumeSpecName "pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.475426 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfrlz\" (UniqueName: \"kubernetes.io/projected/c5917d38-82ff-4f41-ae15-61aca63e378c-kube-api-access-gfrlz\") on node \"crc\" DevicePath \"\"" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.475523 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\") on node \"crc\" " Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.602475 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.616258 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.862516 4945 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.863214 4945 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3") on node "crc" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.884366 4945 reconciler_common.go:293] "Volume detached for volume \"pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c99f1a75-7d47-41b6-87a6-95c7412c9ca3\") on node \"crc\" DevicePath \"\"" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.953666 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:23:52 crc kubenswrapper[4945]: E1206 09:23:52.954345 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:23:52 crc kubenswrapper[4945]: I1206 09:23:52.971681 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5917d38-82ff-4f41-ae15-61aca63e378c" path="/var/lib/kubelet/pods/c5917d38-82ff-4f41-ae15-61aca63e378c/volumes" Dec 06 09:23:53 crc kubenswrapper[4945]: I1206 09:23:53.257791 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 06 09:23:53 crc kubenswrapper[4945]: I1206 09:23:53.258129 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="3cad8b6b-afb4-407a-8bbd-b3e407fb957a" containerName="adoption" containerID="cri-o://8eb065542828012f6bc66c86d09875029a70ae45331820fc410c77a44d997877" gracePeriod=30 Dec 06 09:24:07 crc kubenswrapper[4945]: I1206 09:24:07.954013 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:24:07 crc kubenswrapper[4945]: E1206 09:24:07.954807 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:24:19 crc kubenswrapper[4945]: I1206 09:24:19.952792 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:24:19 crc kubenswrapper[4945]: E1206 09:24:19.953568 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.546076 4945 generic.go:334] "Generic (PLEG): container finished" podID="3cad8b6b-afb4-407a-8bbd-b3e407fb957a" containerID="8eb065542828012f6bc66c86d09875029a70ae45331820fc410c77a44d997877" exitCode=137 Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.546183 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3cad8b6b-afb4-407a-8bbd-b3e407fb957a","Type":"ContainerDied","Data":"8eb065542828012f6bc66c86d09875029a70ae45331820fc410c77a44d997877"} Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.766217 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.835779 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkd85\" (UniqueName: \"kubernetes.io/projected/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-kube-api-access-nkd85\") pod \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.835842 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-ovn-data-cert\") pod \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.836386 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\") pod \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\" (UID: \"3cad8b6b-afb4-407a-8bbd-b3e407fb957a\") " Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.842498 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-kube-api-access-nkd85" (OuterVolumeSpecName: "kube-api-access-nkd85") pod "3cad8b6b-afb4-407a-8bbd-b3e407fb957a" (UID: "3cad8b6b-afb4-407a-8bbd-b3e407fb957a"). InnerVolumeSpecName "kube-api-access-nkd85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.844136 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "3cad8b6b-afb4-407a-8bbd-b3e407fb957a" (UID: "3cad8b6b-afb4-407a-8bbd-b3e407fb957a"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.859458 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9" (OuterVolumeSpecName: "ovn-data") pod "3cad8b6b-afb4-407a-8bbd-b3e407fb957a" (UID: "3cad8b6b-afb4-407a-8bbd-b3e407fb957a"). InnerVolumeSpecName "pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.938771 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkd85\" (UniqueName: \"kubernetes.io/projected/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-kube-api-access-nkd85\") on node \"crc\" DevicePath \"\"" Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.938814 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3cad8b6b-afb4-407a-8bbd-b3e407fb957a-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.938864 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\") on node \"crc\" " Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.963941 4945 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 06 09:24:23 crc kubenswrapper[4945]: I1206 09:24:23.964102 4945 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9") on node "crc" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.040418 4945 reconciler_common.go:293] "Volume detached for volume \"pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1be4b1e2-c756-4712-ad55-f509d9c91ba9\") on node \"crc\" DevicePath \"\"" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.231780 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t6f59"] Dec 06 09:24:24 crc kubenswrapper[4945]: E1206 09:24:24.232193 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5917d38-82ff-4f41-ae15-61aca63e378c" containerName="adoption" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.232207 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5917d38-82ff-4f41-ae15-61aca63e378c" containerName="adoption" Dec 06 09:24:24 crc kubenswrapper[4945]: E1206 09:24:24.232233 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b79a891-f567-4de9-be19-fae05a6722be" containerName="extract-utilities" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.232240 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b79a891-f567-4de9-be19-fae05a6722be" containerName="extract-utilities" Dec 06 09:24:24 crc kubenswrapper[4945]: E1206 09:24:24.232253 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b79a891-f567-4de9-be19-fae05a6722be" containerName="registry-server" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.232259 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b79a891-f567-4de9-be19-fae05a6722be" containerName="registry-server" Dec 06 09:24:24 crc kubenswrapper[4945]: E1206 09:24:24.232289 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cad8b6b-afb4-407a-8bbd-b3e407fb957a" containerName="adoption" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.232295 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cad8b6b-afb4-407a-8bbd-b3e407fb957a" containerName="adoption" Dec 06 09:24:24 crc kubenswrapper[4945]: E1206 09:24:24.232309 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b79a891-f567-4de9-be19-fae05a6722be" containerName="extract-content" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.232315 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b79a891-f567-4de9-be19-fae05a6722be" containerName="extract-content" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.232517 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b79a891-f567-4de9-be19-fae05a6722be" containerName="registry-server" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.232530 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5917d38-82ff-4f41-ae15-61aca63e378c" containerName="adoption" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.232547 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cad8b6b-afb4-407a-8bbd-b3e407fb957a" containerName="adoption" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.234024 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.255341 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6f59"] Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.345393 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq8qc\" (UniqueName: \"kubernetes.io/projected/c4791929-0eed-46ef-bfa8-031eceb942af-kube-api-access-dq8qc\") pod \"redhat-marketplace-t6f59\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.345632 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-catalog-content\") pod \"redhat-marketplace-t6f59\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.345667 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-utilities\") pod \"redhat-marketplace-t6f59\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.447872 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq8qc\" (UniqueName: \"kubernetes.io/projected/c4791929-0eed-46ef-bfa8-031eceb942af-kube-api-access-dq8qc\") pod \"redhat-marketplace-t6f59\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.448118 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-catalog-content\") pod \"redhat-marketplace-t6f59\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.448166 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-utilities\") pod \"redhat-marketplace-t6f59\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.448661 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-catalog-content\") pod \"redhat-marketplace-t6f59\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.448802 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-utilities\") pod \"redhat-marketplace-t6f59\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.466530 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq8qc\" (UniqueName: \"kubernetes.io/projected/c4791929-0eed-46ef-bfa8-031eceb942af-kube-api-access-dq8qc\") pod \"redhat-marketplace-t6f59\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.551944 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.557515 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3cad8b6b-afb4-407a-8bbd-b3e407fb957a","Type":"ContainerDied","Data":"3182645f89f4ac6300b6379cd7cb04dd80a7eb1f86fc30559ee0cba5d6e7b491"} Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.557571 4945 scope.go:117] "RemoveContainer" containerID="8eb065542828012f6bc66c86d09875029a70ae45331820fc410c77a44d997877" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.557583 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.592170 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.605963 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Dec 06 09:24:24 crc kubenswrapper[4945]: I1206 09:24:24.964401 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cad8b6b-afb4-407a-8bbd-b3e407fb957a" path="/var/lib/kubelet/pods/3cad8b6b-afb4-407a-8bbd-b3e407fb957a/volumes" Dec 06 09:24:25 crc kubenswrapper[4945]: I1206 09:24:25.108715 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6f59"] Dec 06 09:24:25 crc kubenswrapper[4945]: I1206 09:24:25.575718 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6f59" event={"ID":"c4791929-0eed-46ef-bfa8-031eceb942af","Type":"ContainerStarted","Data":"536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952"} Dec 06 09:24:25 crc kubenswrapper[4945]: I1206 09:24:25.575774 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6f59" event={"ID":"c4791929-0eed-46ef-bfa8-031eceb942af","Type":"ContainerStarted","Data":"83ceb9840b966d59556c15430ac85548afcda81c335d8b161bef0bb1073bac20"} Dec 06 09:24:26 crc kubenswrapper[4945]: I1206 09:24:26.588477 4945 generic.go:334] "Generic (PLEG): container finished" podID="c4791929-0eed-46ef-bfa8-031eceb942af" containerID="536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952" exitCode=0 Dec 06 09:24:26 crc kubenswrapper[4945]: I1206 09:24:26.588586 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6f59" event={"ID":"c4791929-0eed-46ef-bfa8-031eceb942af","Type":"ContainerDied","Data":"536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952"} Dec 06 09:24:27 crc kubenswrapper[4945]: I1206 09:24:27.600906 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6f59" event={"ID":"c4791929-0eed-46ef-bfa8-031eceb942af","Type":"ContainerStarted","Data":"3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771"} Dec 06 09:24:28 crc kubenswrapper[4945]: I1206 09:24:28.611306 4945 generic.go:334] "Generic (PLEG): container finished" podID="c4791929-0eed-46ef-bfa8-031eceb942af" containerID="3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771" exitCode=0 Dec 06 09:24:28 crc kubenswrapper[4945]: I1206 09:24:28.611347 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6f59" event={"ID":"c4791929-0eed-46ef-bfa8-031eceb942af","Type":"ContainerDied","Data":"3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771"} Dec 06 09:24:29 crc kubenswrapper[4945]: I1206 09:24:29.622918 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6f59" event={"ID":"c4791929-0eed-46ef-bfa8-031eceb942af","Type":"ContainerStarted","Data":"f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb"} Dec 06 09:24:32 crc kubenswrapper[4945]: I1206 09:24:32.953463 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:24:32 crc kubenswrapper[4945]: E1206 09:24:32.955161 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:24:34 crc kubenswrapper[4945]: I1206 09:24:34.553132 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:34 crc kubenswrapper[4945]: I1206 09:24:34.553620 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:34 crc kubenswrapper[4945]: I1206 09:24:34.600834 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:34 crc kubenswrapper[4945]: I1206 09:24:34.619747 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t6f59" podStartSLOduration=7.187989534 podStartE2EDuration="10.619725757s" podCreationTimestamp="2025-12-06 09:24:24 +0000 UTC" firstStartedPulling="2025-12-06 09:24:25.579646873 +0000 UTC m=+9079.034507917" lastFinishedPulling="2025-12-06 09:24:29.011383096 +0000 UTC m=+9082.466244140" observedRunningTime="2025-12-06 09:24:29.64832311 +0000 UTC m=+9083.103184164" watchObservedRunningTime="2025-12-06 09:24:34.619725757 +0000 UTC m=+9088.074586801" Dec 06 09:24:34 crc kubenswrapper[4945]: I1206 09:24:34.727887 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:34 crc kubenswrapper[4945]: I1206 09:24:34.836484 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6f59"] Dec 06 09:24:36 crc kubenswrapper[4945]: I1206 09:24:36.695119 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t6f59" podUID="c4791929-0eed-46ef-bfa8-031eceb942af" containerName="registry-server" containerID="cri-o://f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb" gracePeriod=2 Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.226050 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.350981 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-catalog-content\") pod \"c4791929-0eed-46ef-bfa8-031eceb942af\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.351184 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq8qc\" (UniqueName: \"kubernetes.io/projected/c4791929-0eed-46ef-bfa8-031eceb942af-kube-api-access-dq8qc\") pod \"c4791929-0eed-46ef-bfa8-031eceb942af\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.351245 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-utilities\") pod \"c4791929-0eed-46ef-bfa8-031eceb942af\" (UID: \"c4791929-0eed-46ef-bfa8-031eceb942af\") " Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.352162 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-utilities" (OuterVolumeSpecName: "utilities") pod "c4791929-0eed-46ef-bfa8-031eceb942af" (UID: "c4791929-0eed-46ef-bfa8-031eceb942af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.371555 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4791929-0eed-46ef-bfa8-031eceb942af" (UID: "c4791929-0eed-46ef-bfa8-031eceb942af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.454054 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.454081 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4791929-0eed-46ef-bfa8-031eceb942af-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.719992 4945 generic.go:334] "Generic (PLEG): container finished" podID="c4791929-0eed-46ef-bfa8-031eceb942af" containerID="f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb" exitCode=0 Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.720077 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6f59" event={"ID":"c4791929-0eed-46ef-bfa8-031eceb942af","Type":"ContainerDied","Data":"f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb"} Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.720140 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6f59" event={"ID":"c4791929-0eed-46ef-bfa8-031eceb942af","Type":"ContainerDied","Data":"83ceb9840b966d59556c15430ac85548afcda81c335d8b161bef0bb1073bac20"} Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.720169 4945 scope.go:117] "RemoveContainer" containerID="f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb" Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.720555 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t6f59" Dec 06 09:24:37 crc kubenswrapper[4945]: I1206 09:24:37.750149 4945 scope.go:117] "RemoveContainer" containerID="3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771" Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.039231 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4791929-0eed-46ef-bfa8-031eceb942af-kube-api-access-dq8qc" (OuterVolumeSpecName: "kube-api-access-dq8qc") pod "c4791929-0eed-46ef-bfa8-031eceb942af" (UID: "c4791929-0eed-46ef-bfa8-031eceb942af"). InnerVolumeSpecName "kube-api-access-dq8qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.062625 4945 scope.go:117] "RemoveContainer" containerID="536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952" Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.066997 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq8qc\" (UniqueName: \"kubernetes.io/projected/c4791929-0eed-46ef-bfa8-031eceb942af-kube-api-access-dq8qc\") on node \"crc\" DevicePath \"\"" Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.236470 4945 scope.go:117] "RemoveContainer" containerID="f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb" Dec 06 09:24:38 crc kubenswrapper[4945]: E1206 09:24:38.236858 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb\": container with ID starting with f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb not found: ID does not exist" containerID="f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb" Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.236901 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb"} err="failed to get container status \"f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb\": rpc error: code = NotFound desc = could not find container \"f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb\": container with ID starting with f09e81103592330a7c87ce882a97d852211e971e939fa58541af6c64ed87cbeb not found: ID does not exist" Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.236928 4945 scope.go:117] "RemoveContainer" containerID="3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771" Dec 06 09:24:38 crc kubenswrapper[4945]: E1206 09:24:38.237201 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771\": container with ID starting with 3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771 not found: ID does not exist" containerID="3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771" Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.237227 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771"} err="failed to get container status \"3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771\": rpc error: code = NotFound desc = could not find container \"3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771\": container with ID starting with 3b7888e3b6dbe602f5c87600e38797f31d80b5cd3011cc7aaf0f51e582baf771 not found: ID does not exist" Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.237244 4945 scope.go:117] "RemoveContainer" containerID="536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952" Dec 06 09:24:38 crc kubenswrapper[4945]: E1206 09:24:38.237516 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952\": container with ID starting with 536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952 not found: ID does not exist" containerID="536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952" Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.237540 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952"} err="failed to get container status \"536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952\": rpc error: code = NotFound desc = could not find container \"536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952\": container with ID starting with 536cb951a5022f1067251f61ea09f100dba45c93a674eccfe65c6331e427e952 not found: ID does not exist" Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.361830 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6f59"] Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.372078 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6f59"] Dec 06 09:24:38 crc kubenswrapper[4945]: I1206 09:24:38.965367 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4791929-0eed-46ef-bfa8-031eceb942af" path="/var/lib/kubelet/pods/c4791929-0eed-46ef-bfa8-031eceb942af/volumes" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.947068 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 09:24:44 crc kubenswrapper[4945]: E1206 09:24:44.948390 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4791929-0eed-46ef-bfa8-031eceb942af" containerName="extract-utilities" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.948410 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4791929-0eed-46ef-bfa8-031eceb942af" containerName="extract-utilities" Dec 06 09:24:44 crc kubenswrapper[4945]: E1206 09:24:44.948453 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4791929-0eed-46ef-bfa8-031eceb942af" containerName="registry-server" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.948463 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4791929-0eed-46ef-bfa8-031eceb942af" containerName="registry-server" Dec 06 09:24:44 crc kubenswrapper[4945]: E1206 09:24:44.948494 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4791929-0eed-46ef-bfa8-031eceb942af" containerName="extract-content" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.948502 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4791929-0eed-46ef-bfa8-031eceb942af" containerName="extract-content" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.948950 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4791929-0eed-46ef-bfa8-031eceb942af" containerName="registry-server" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.949952 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.952516 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.953010 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.953144 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.953297 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gxzcx" Dec 06 09:24:44 crc kubenswrapper[4945]: I1206 09:24:44.974895 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.002438 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.002598 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.002668 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhsjq\" (UniqueName: \"kubernetes.io/projected/1000996d-89a3-43f6-bffc-03fd64593568-kube-api-access-lhsjq\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.002942 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.003012 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.003107 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-config-data\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.003145 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.003193 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.003295 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.104897 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.104945 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.104984 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-config-data\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.105013 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.105045 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.105083 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.105144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.105163 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.105194 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhsjq\" (UniqueName: \"kubernetes.io/projected/1000996d-89a3-43f6-bffc-03fd64593568-kube-api-access-lhsjq\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.105894 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.106014 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.106235 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.108250 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.108387 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-config-data\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.110763 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.112107 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.113974 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.123269 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhsjq\" (UniqueName: \"kubernetes.io/projected/1000996d-89a3-43f6-bffc-03fd64593568-kube-api-access-lhsjq\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.141624 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.278330 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.724303 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 09:24:45 crc kubenswrapper[4945]: I1206 09:24:45.793714 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1000996d-89a3-43f6-bffc-03fd64593568","Type":"ContainerStarted","Data":"785467adddf6845ec72e009f988d0858cf35b599dc02309bcabe9976ef4c8e68"} Dec 06 09:24:46 crc kubenswrapper[4945]: I1206 09:24:46.975312 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:24:46 crc kubenswrapper[4945]: E1206 09:24:46.976007 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:25:02 crc kubenswrapper[4945]: I1206 09:25:02.953434 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:25:02 crc kubenswrapper[4945]: E1206 09:25:02.954302 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:25:15 crc kubenswrapper[4945]: I1206 09:25:15.955248 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:25:15 crc kubenswrapper[4945]: E1206 09:25:15.956245 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:25:30 crc kubenswrapper[4945]: I1206 09:25:30.953901 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:25:30 crc kubenswrapper[4945]: E1206 09:25:30.954716 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:25:32 crc kubenswrapper[4945]: E1206 09:25:32.152039 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:c3923531bcda0b0811b2d5053f189beb" Dec 06 09:25:32 crc kubenswrapper[4945]: E1206 09:25:32.152094 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:c3923531bcda0b0811b2d5053f189beb" Dec 06 09:25:32 crc kubenswrapper[4945]: E1206 09:25:32.152228 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:c3923531bcda0b0811b2d5053f189beb,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lhsjq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(1000996d-89a3-43f6-bffc-03fd64593568): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 09:25:32 crc kubenswrapper[4945]: E1206 09:25:32.153509 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="1000996d-89a3-43f6-bffc-03fd64593568" Dec 06 09:25:32 crc kubenswrapper[4945]: E1206 09:25:32.309488 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:c3923531bcda0b0811b2d5053f189beb\\\"\"" pod="openstack/tempest-tests-tempest" podUID="1000996d-89a3-43f6-bffc-03fd64593568" Dec 06 09:25:41 crc kubenswrapper[4945]: I1206 09:25:41.953760 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:25:41 crc kubenswrapper[4945]: E1206 09:25:41.954657 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:25:45 crc kubenswrapper[4945]: I1206 09:25:45.171889 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.331153 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8s4jw"] Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.335131 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.342825 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8s4jw"] Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.429614 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-utilities\") pod \"redhat-operators-8s4jw\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.430000 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-catalog-content\") pod \"redhat-operators-8s4jw\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.430040 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcpr8\" (UniqueName: \"kubernetes.io/projected/223f191e-94fb-4662-abcf-4d81b5ee08ea-kube-api-access-lcpr8\") pod \"redhat-operators-8s4jw\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.531589 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-utilities\") pod \"redhat-operators-8s4jw\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.531960 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-catalog-content\") pod \"redhat-operators-8s4jw\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.532038 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-utilities\") pod \"redhat-operators-8s4jw\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.532057 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcpr8\" (UniqueName: \"kubernetes.io/projected/223f191e-94fb-4662-abcf-4d81b5ee08ea-kube-api-access-lcpr8\") pod \"redhat-operators-8s4jw\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.532340 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-catalog-content\") pod \"redhat-operators-8s4jw\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.551474 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcpr8\" (UniqueName: \"kubernetes.io/projected/223f191e-94fb-4662-abcf-4d81b5ee08ea-kube-api-access-lcpr8\") pod \"redhat-operators-8s4jw\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:46 crc kubenswrapper[4945]: I1206 09:25:46.661786 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:25:47 crc kubenswrapper[4945]: I1206 09:25:47.190364 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8s4jw"] Dec 06 09:25:47 crc kubenswrapper[4945]: W1206 09:25:47.193421 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod223f191e_94fb_4662_abcf_4d81b5ee08ea.slice/crio-403a443d3420a778c9775da3f164c50996d2da6f0eebcc74c00658de876dbcaf WatchSource:0}: Error finding container 403a443d3420a778c9775da3f164c50996d2da6f0eebcc74c00658de876dbcaf: Status 404 returned error can't find the container with id 403a443d3420a778c9775da3f164c50996d2da6f0eebcc74c00658de876dbcaf Dec 06 09:25:47 crc kubenswrapper[4945]: I1206 09:25:47.442983 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1000996d-89a3-43f6-bffc-03fd64593568","Type":"ContainerStarted","Data":"141ee0777a89d8c08f7ff6884c81911edefb7e920eacd6752a41565cdb06fbe5"} Dec 06 09:25:47 crc kubenswrapper[4945]: I1206 09:25:47.446041 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8s4jw" event={"ID":"223f191e-94fb-4662-abcf-4d81b5ee08ea","Type":"ContainerStarted","Data":"c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8"} Dec 06 09:25:47 crc kubenswrapper[4945]: I1206 09:25:47.446093 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8s4jw" event={"ID":"223f191e-94fb-4662-abcf-4d81b5ee08ea","Type":"ContainerStarted","Data":"403a443d3420a778c9775da3f164c50996d2da6f0eebcc74c00658de876dbcaf"} Dec 06 09:25:47 crc kubenswrapper[4945]: I1206 09:25:47.479403 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=5.058590941 podStartE2EDuration="1m4.479378703s" podCreationTimestamp="2025-12-06 09:24:43 +0000 UTC" firstStartedPulling="2025-12-06 09:24:45.730571117 +0000 UTC m=+9099.185432161" lastFinishedPulling="2025-12-06 09:25:45.151358879 +0000 UTC m=+9158.606219923" observedRunningTime="2025-12-06 09:25:47.466065417 +0000 UTC m=+9160.920926461" watchObservedRunningTime="2025-12-06 09:25:47.479378703 +0000 UTC m=+9160.934239737" Dec 06 09:25:48 crc kubenswrapper[4945]: I1206 09:25:48.458772 4945 generic.go:334] "Generic (PLEG): container finished" podID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerID="c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8" exitCode=0 Dec 06 09:25:48 crc kubenswrapper[4945]: I1206 09:25:48.458824 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8s4jw" event={"ID":"223f191e-94fb-4662-abcf-4d81b5ee08ea","Type":"ContainerDied","Data":"c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8"} Dec 06 09:25:49 crc kubenswrapper[4945]: I1206 09:25:49.471489 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8s4jw" event={"ID":"223f191e-94fb-4662-abcf-4d81b5ee08ea","Type":"ContainerStarted","Data":"e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef"} Dec 06 09:25:52 crc kubenswrapper[4945]: I1206 09:25:52.954478 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:25:52 crc kubenswrapper[4945]: E1206 09:25:52.955016 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:25:55 crc kubenswrapper[4945]: I1206 09:25:55.543201 4945 generic.go:334] "Generic (PLEG): container finished" podID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerID="e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef" exitCode=0 Dec 06 09:25:55 crc kubenswrapper[4945]: I1206 09:25:55.543293 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8s4jw" event={"ID":"223f191e-94fb-4662-abcf-4d81b5ee08ea","Type":"ContainerDied","Data":"e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef"} Dec 06 09:25:57 crc kubenswrapper[4945]: I1206 09:25:57.563496 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8s4jw" event={"ID":"223f191e-94fb-4662-abcf-4d81b5ee08ea","Type":"ContainerStarted","Data":"c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24"} Dec 06 09:25:57 crc kubenswrapper[4945]: I1206 09:25:57.587764 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8s4jw" podStartSLOduration=3.003054798 podStartE2EDuration="11.58774799s" podCreationTimestamp="2025-12-06 09:25:46 +0000 UTC" firstStartedPulling="2025-12-06 09:25:48.461060821 +0000 UTC m=+9161.915921875" lastFinishedPulling="2025-12-06 09:25:57.045754033 +0000 UTC m=+9170.500615067" observedRunningTime="2025-12-06 09:25:57.584600066 +0000 UTC m=+9171.039461110" watchObservedRunningTime="2025-12-06 09:25:57.58774799 +0000 UTC m=+9171.042609034" Dec 06 09:26:04 crc kubenswrapper[4945]: I1206 09:26:04.953364 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:26:04 crc kubenswrapper[4945]: E1206 09:26:04.954170 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:26:06 crc kubenswrapper[4945]: I1206 09:26:06.662228 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:26:06 crc kubenswrapper[4945]: I1206 09:26:06.662554 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:26:06 crc kubenswrapper[4945]: I1206 09:26:06.709259 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:26:07 crc kubenswrapper[4945]: I1206 09:26:07.693269 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:26:07 crc kubenswrapper[4945]: I1206 09:26:07.761133 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8s4jw"] Dec 06 09:26:09 crc kubenswrapper[4945]: I1206 09:26:09.663577 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8s4jw" podUID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerName="registry-server" containerID="cri-o://c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24" gracePeriod=2 Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.396730 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.515763 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-utilities\") pod \"223f191e-94fb-4662-abcf-4d81b5ee08ea\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.516143 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcpr8\" (UniqueName: \"kubernetes.io/projected/223f191e-94fb-4662-abcf-4d81b5ee08ea-kube-api-access-lcpr8\") pod \"223f191e-94fb-4662-abcf-4d81b5ee08ea\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.516232 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-catalog-content\") pod \"223f191e-94fb-4662-abcf-4d81b5ee08ea\" (UID: \"223f191e-94fb-4662-abcf-4d81b5ee08ea\") " Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.516580 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-utilities" (OuterVolumeSpecName: "utilities") pod "223f191e-94fb-4662-abcf-4d81b5ee08ea" (UID: "223f191e-94fb-4662-abcf-4d81b5ee08ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.516726 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.521666 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/223f191e-94fb-4662-abcf-4d81b5ee08ea-kube-api-access-lcpr8" (OuterVolumeSpecName: "kube-api-access-lcpr8") pod "223f191e-94fb-4662-abcf-4d81b5ee08ea" (UID: "223f191e-94fb-4662-abcf-4d81b5ee08ea"). InnerVolumeSpecName "kube-api-access-lcpr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.618574 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcpr8\" (UniqueName: \"kubernetes.io/projected/223f191e-94fb-4662-abcf-4d81b5ee08ea-kube-api-access-lcpr8\") on node \"crc\" DevicePath \"\"" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.634548 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "223f191e-94fb-4662-abcf-4d81b5ee08ea" (UID: "223f191e-94fb-4662-abcf-4d81b5ee08ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.674852 4945 generic.go:334] "Generic (PLEG): container finished" podID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerID="c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24" exitCode=0 Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.674899 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8s4jw" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.674902 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8s4jw" event={"ID":"223f191e-94fb-4662-abcf-4d81b5ee08ea","Type":"ContainerDied","Data":"c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24"} Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.674931 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8s4jw" event={"ID":"223f191e-94fb-4662-abcf-4d81b5ee08ea","Type":"ContainerDied","Data":"403a443d3420a778c9775da3f164c50996d2da6f0eebcc74c00658de876dbcaf"} Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.674953 4945 scope.go:117] "RemoveContainer" containerID="c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.703507 4945 scope.go:117] "RemoveContainer" containerID="e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.713124 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8s4jw"] Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.721060 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/223f191e-94fb-4662-abcf-4d81b5ee08ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.723601 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8s4jw"] Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.736413 4945 scope.go:117] "RemoveContainer" containerID="c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.775016 4945 scope.go:117] "RemoveContainer" containerID="c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24" Dec 06 09:26:10 crc kubenswrapper[4945]: E1206 09:26:10.775596 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24\": container with ID starting with c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24 not found: ID does not exist" containerID="c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.775632 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24"} err="failed to get container status \"c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24\": rpc error: code = NotFound desc = could not find container \"c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24\": container with ID starting with c612267b4b04581c21274a0b7b79442c3245c6cb2033612c2278cff8aa39af24 not found: ID does not exist" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.775654 4945 scope.go:117] "RemoveContainer" containerID="e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef" Dec 06 09:26:10 crc kubenswrapper[4945]: E1206 09:26:10.776025 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef\": container with ID starting with e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef not found: ID does not exist" containerID="e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.776092 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef"} err="failed to get container status \"e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef\": rpc error: code = NotFound desc = could not find container \"e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef\": container with ID starting with e9a79e8d8c54b6b6aeedb0705a26707c1de5ca532f1e99b8e06cd04e589a65ef not found: ID does not exist" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.776108 4945 scope.go:117] "RemoveContainer" containerID="c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8" Dec 06 09:26:10 crc kubenswrapper[4945]: E1206 09:26:10.776541 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8\": container with ID starting with c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8 not found: ID does not exist" containerID="c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.776593 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8"} err="failed to get container status \"c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8\": rpc error: code = NotFound desc = could not find container \"c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8\": container with ID starting with c26542b4534e88e6b6f45bea74105d65772175c0758f2d7ad82f6e4e880b4ba8 not found: ID does not exist" Dec 06 09:26:10 crc kubenswrapper[4945]: I1206 09:26:10.966652 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="223f191e-94fb-4662-abcf-4d81b5ee08ea" path="/var/lib/kubelet/pods/223f191e-94fb-4662-abcf-4d81b5ee08ea/volumes" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.551973 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x4lhb"] Dec 06 09:26:14 crc kubenswrapper[4945]: E1206 09:26:14.553182 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerName="registry-server" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.553196 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerName="registry-server" Dec 06 09:26:14 crc kubenswrapper[4945]: E1206 09:26:14.553209 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerName="extract-content" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.553218 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerName="extract-content" Dec 06 09:26:14 crc kubenswrapper[4945]: E1206 09:26:14.553252 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerName="extract-utilities" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.553259 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerName="extract-utilities" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.553518 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="223f191e-94fb-4662-abcf-4d81b5ee08ea" containerName="registry-server" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.555241 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.583093 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x4lhb"] Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.604949 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-catalog-content\") pod \"community-operators-x4lhb\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.606353 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9bs8\" (UniqueName: \"kubernetes.io/projected/e03a64d2-d54f-4ff2-9088-53c71671ce9f-kube-api-access-d9bs8\") pod \"community-operators-x4lhb\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.607971 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-utilities\") pod \"community-operators-x4lhb\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.710259 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-utilities\") pod \"community-operators-x4lhb\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.710393 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-catalog-content\") pod \"community-operators-x4lhb\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.710456 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9bs8\" (UniqueName: \"kubernetes.io/projected/e03a64d2-d54f-4ff2-9088-53c71671ce9f-kube-api-access-d9bs8\") pod \"community-operators-x4lhb\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.710898 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-catalog-content\") pod \"community-operators-x4lhb\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:14 crc kubenswrapper[4945]: I1206 09:26:14.710908 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-utilities\") pod \"community-operators-x4lhb\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:17 crc kubenswrapper[4945]: I1206 09:26:17.041300 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9bs8\" (UniqueName: \"kubernetes.io/projected/e03a64d2-d54f-4ff2-9088-53c71671ce9f-kube-api-access-d9bs8\") pod \"community-operators-x4lhb\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:17 crc kubenswrapper[4945]: I1206 09:26:17.153198 4945 trace.go:236] Trace[1034891190]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-cell1-server-0" (06-Dec-2025 09:26:14.909) (total time: 2243ms): Dec 06 09:26:17 crc kubenswrapper[4945]: Trace[1034891190]: [2.24334081s] [2.24334081s] END Dec 06 09:26:17 crc kubenswrapper[4945]: I1206 09:26:17.305178 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:17 crc kubenswrapper[4945]: I1206 09:26:17.811807 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x4lhb"] Dec 06 09:26:18 crc kubenswrapper[4945]: I1206 09:26:18.756721 4945 generic.go:334] "Generic (PLEG): container finished" podID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerID="1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c" exitCode=0 Dec 06 09:26:18 crc kubenswrapper[4945]: I1206 09:26:18.756842 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4lhb" event={"ID":"e03a64d2-d54f-4ff2-9088-53c71671ce9f","Type":"ContainerDied","Data":"1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c"} Dec 06 09:26:18 crc kubenswrapper[4945]: I1206 09:26:18.757123 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4lhb" event={"ID":"e03a64d2-d54f-4ff2-9088-53c71671ce9f","Type":"ContainerStarted","Data":"7e1c4b95c99d2b5cd806c0bf2168518fda418c17487bd4b732c32c29e416dbd0"} Dec 06 09:26:19 crc kubenswrapper[4945]: I1206 09:26:19.768034 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4lhb" event={"ID":"e03a64d2-d54f-4ff2-9088-53c71671ce9f","Type":"ContainerStarted","Data":"a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4"} Dec 06 09:26:19 crc kubenswrapper[4945]: I1206 09:26:19.954882 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:26:20 crc kubenswrapper[4945]: I1206 09:26:20.782663 4945 generic.go:334] "Generic (PLEG): container finished" podID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerID="a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4" exitCode=0 Dec 06 09:26:20 crc kubenswrapper[4945]: I1206 09:26:20.782852 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4lhb" event={"ID":"e03a64d2-d54f-4ff2-9088-53c71671ce9f","Type":"ContainerDied","Data":"a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4"} Dec 06 09:26:20 crc kubenswrapper[4945]: I1206 09:26:20.794830 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"46722cb4b2f771a8c019f69da4a13a14521d53249d80f3919d51feb2c76b849a"} Dec 06 09:26:21 crc kubenswrapper[4945]: I1206 09:26:21.811622 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4lhb" event={"ID":"e03a64d2-d54f-4ff2-9088-53c71671ce9f","Type":"ContainerStarted","Data":"d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f"} Dec 06 09:26:21 crc kubenswrapper[4945]: I1206 09:26:21.845955 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x4lhb" podStartSLOduration=5.129792787 podStartE2EDuration="7.845936324s" podCreationTimestamp="2025-12-06 09:26:14 +0000 UTC" firstStartedPulling="2025-12-06 09:26:18.758718418 +0000 UTC m=+9192.213579462" lastFinishedPulling="2025-12-06 09:26:21.474861955 +0000 UTC m=+9194.929722999" observedRunningTime="2025-12-06 09:26:21.831891998 +0000 UTC m=+9195.286753042" watchObservedRunningTime="2025-12-06 09:26:21.845936324 +0000 UTC m=+9195.300797368" Dec 06 09:26:27 crc kubenswrapper[4945]: I1206 09:26:27.305596 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:27 crc kubenswrapper[4945]: I1206 09:26:27.306159 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:27 crc kubenswrapper[4945]: I1206 09:26:27.356078 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:27 crc kubenswrapper[4945]: I1206 09:26:27.919151 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:27 crc kubenswrapper[4945]: I1206 09:26:27.968230 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x4lhb"] Dec 06 09:26:29 crc kubenswrapper[4945]: I1206 09:26:29.896191 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x4lhb" podUID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerName="registry-server" containerID="cri-o://d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f" gracePeriod=2 Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.478682 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.633168 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9bs8\" (UniqueName: \"kubernetes.io/projected/e03a64d2-d54f-4ff2-9088-53c71671ce9f-kube-api-access-d9bs8\") pod \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.633389 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-utilities\") pod \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.633416 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-catalog-content\") pod \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\" (UID: \"e03a64d2-d54f-4ff2-9088-53c71671ce9f\") " Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.635028 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-utilities" (OuterVolumeSpecName: "utilities") pod "e03a64d2-d54f-4ff2-9088-53c71671ce9f" (UID: "e03a64d2-d54f-4ff2-9088-53c71671ce9f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.650510 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e03a64d2-d54f-4ff2-9088-53c71671ce9f-kube-api-access-d9bs8" (OuterVolumeSpecName: "kube-api-access-d9bs8") pod "e03a64d2-d54f-4ff2-9088-53c71671ce9f" (UID: "e03a64d2-d54f-4ff2-9088-53c71671ce9f"). InnerVolumeSpecName "kube-api-access-d9bs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.698215 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e03a64d2-d54f-4ff2-9088-53c71671ce9f" (UID: "e03a64d2-d54f-4ff2-9088-53c71671ce9f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.764158 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.764196 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e03a64d2-d54f-4ff2-9088-53c71671ce9f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.764210 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9bs8\" (UniqueName: \"kubernetes.io/projected/e03a64d2-d54f-4ff2-9088-53c71671ce9f-kube-api-access-d9bs8\") on node \"crc\" DevicePath \"\"" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.913558 4945 generic.go:334] "Generic (PLEG): container finished" podID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerID="d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f" exitCode=0 Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.913599 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4lhb" event={"ID":"e03a64d2-d54f-4ff2-9088-53c71671ce9f","Type":"ContainerDied","Data":"d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f"} Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.913791 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4lhb" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.913813 4945 scope.go:117] "RemoveContainer" containerID="d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.913802 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4lhb" event={"ID":"e03a64d2-d54f-4ff2-9088-53c71671ce9f","Type":"ContainerDied","Data":"7e1c4b95c99d2b5cd806c0bf2168518fda418c17487bd4b732c32c29e416dbd0"} Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.945399 4945 scope.go:117] "RemoveContainer" containerID="a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.970218 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x4lhb"] Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.980397 4945 scope.go:117] "RemoveContainer" containerID="1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c" Dec 06 09:26:30 crc kubenswrapper[4945]: I1206 09:26:30.983633 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x4lhb"] Dec 06 09:26:31 crc kubenswrapper[4945]: I1206 09:26:31.021239 4945 scope.go:117] "RemoveContainer" containerID="d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f" Dec 06 09:26:31 crc kubenswrapper[4945]: E1206 09:26:31.021722 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f\": container with ID starting with d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f not found: ID does not exist" containerID="d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f" Dec 06 09:26:31 crc kubenswrapper[4945]: I1206 09:26:31.021775 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f"} err="failed to get container status \"d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f\": rpc error: code = NotFound desc = could not find container \"d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f\": container with ID starting with d63c8633561a49d2e617b4954236933b0efb6b9de2925cf720ecfec0d20c304f not found: ID does not exist" Dec 06 09:26:31 crc kubenswrapper[4945]: I1206 09:26:31.021810 4945 scope.go:117] "RemoveContainer" containerID="a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4" Dec 06 09:26:31 crc kubenswrapper[4945]: E1206 09:26:31.022193 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4\": container with ID starting with a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4 not found: ID does not exist" containerID="a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4" Dec 06 09:26:31 crc kubenswrapper[4945]: I1206 09:26:31.022223 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4"} err="failed to get container status \"a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4\": rpc error: code = NotFound desc = could not find container \"a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4\": container with ID starting with a500f500f9bbb52e8e0f2839e100dc24b8da37e53157b573442d7f8eacab0eb4 not found: ID does not exist" Dec 06 09:26:31 crc kubenswrapper[4945]: I1206 09:26:31.022243 4945 scope.go:117] "RemoveContainer" containerID="1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c" Dec 06 09:26:31 crc kubenswrapper[4945]: E1206 09:26:31.022700 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c\": container with ID starting with 1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c not found: ID does not exist" containerID="1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c" Dec 06 09:26:31 crc kubenswrapper[4945]: I1206 09:26:31.022758 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c"} err="failed to get container status \"1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c\": rpc error: code = NotFound desc = could not find container \"1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c\": container with ID starting with 1773ce6840ad8cde73fc98cadff958424e14696067ca51ca0022c9b7b786690c not found: ID does not exist" Dec 06 09:26:32 crc kubenswrapper[4945]: I1206 09:26:32.968623 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" path="/var/lib/kubelet/pods/e03a64d2-d54f-4ff2-9088-53c71671ce9f/volumes" Dec 06 09:28:38 crc kubenswrapper[4945]: I1206 09:28:38.795651 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:28:38 crc kubenswrapper[4945]: I1206 09:28:38.796194 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:29:08 crc kubenswrapper[4945]: I1206 09:29:08.796027 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:29:08 crc kubenswrapper[4945]: I1206 09:29:08.796662 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:29:38 crc kubenswrapper[4945]: I1206 09:29:38.795391 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:29:38 crc kubenswrapper[4945]: I1206 09:29:38.795943 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:29:38 crc kubenswrapper[4945]: I1206 09:29:38.796004 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 09:29:38 crc kubenswrapper[4945]: I1206 09:29:38.796872 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"46722cb4b2f771a8c019f69da4a13a14521d53249d80f3919d51feb2c76b849a"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 09:29:38 crc kubenswrapper[4945]: I1206 09:29:38.796932 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://46722cb4b2f771a8c019f69da4a13a14521d53249d80f3919d51feb2c76b849a" gracePeriod=600 Dec 06 09:29:39 crc kubenswrapper[4945]: I1206 09:29:39.955564 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="46722cb4b2f771a8c019f69da4a13a14521d53249d80f3919d51feb2c76b849a" exitCode=0 Dec 06 09:29:39 crc kubenswrapper[4945]: I1206 09:29:39.956115 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"46722cb4b2f771a8c019f69da4a13a14521d53249d80f3919d51feb2c76b849a"} Dec 06 09:29:39 crc kubenswrapper[4945]: I1206 09:29:39.956143 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f"} Dec 06 09:29:39 crc kubenswrapper[4945]: I1206 09:29:39.956160 4945 scope.go:117] "RemoveContainer" containerID="e8d316c127c83f8b67a7ae5bda09bef41ff87e87a5ffd8feaae00d7da2eab014" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.149012 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w"] Dec 06 09:30:00 crc kubenswrapper[4945]: E1206 09:30:00.150098 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerName="extract-utilities" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.150114 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerName="extract-utilities" Dec 06 09:30:00 crc kubenswrapper[4945]: E1206 09:30:00.150157 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerName="registry-server" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.150165 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerName="registry-server" Dec 06 09:30:00 crc kubenswrapper[4945]: E1206 09:30:00.150178 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerName="extract-content" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.150186 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerName="extract-content" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.150434 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e03a64d2-d54f-4ff2-9088-53c71671ce9f" containerName="registry-server" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.151257 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.153491 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.159948 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.179668 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w"] Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.197243 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66z5d\" (UniqueName: \"kubernetes.io/projected/41e98db9-5682-4905-8ebe-9910226b8a79-kube-api-access-66z5d\") pod \"collect-profiles-29416890-mz44w\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.197583 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41e98db9-5682-4905-8ebe-9910226b8a79-secret-volume\") pod \"collect-profiles-29416890-mz44w\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.197805 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41e98db9-5682-4905-8ebe-9910226b8a79-config-volume\") pod \"collect-profiles-29416890-mz44w\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.299227 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41e98db9-5682-4905-8ebe-9910226b8a79-secret-volume\") pod \"collect-profiles-29416890-mz44w\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.299334 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41e98db9-5682-4905-8ebe-9910226b8a79-config-volume\") pod \"collect-profiles-29416890-mz44w\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.299360 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66z5d\" (UniqueName: \"kubernetes.io/projected/41e98db9-5682-4905-8ebe-9910226b8a79-kube-api-access-66z5d\") pod \"collect-profiles-29416890-mz44w\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.300310 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41e98db9-5682-4905-8ebe-9910226b8a79-config-volume\") pod \"collect-profiles-29416890-mz44w\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.312197 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41e98db9-5682-4905-8ebe-9910226b8a79-secret-volume\") pod \"collect-profiles-29416890-mz44w\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.327860 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66z5d\" (UniqueName: \"kubernetes.io/projected/41e98db9-5682-4905-8ebe-9910226b8a79-kube-api-access-66z5d\") pod \"collect-profiles-29416890-mz44w\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:00 crc kubenswrapper[4945]: I1206 09:30:00.474691 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:01 crc kubenswrapper[4945]: I1206 09:30:01.026087 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w"] Dec 06 09:30:01 crc kubenswrapper[4945]: I1206 09:30:01.190301 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" event={"ID":"41e98db9-5682-4905-8ebe-9910226b8a79","Type":"ContainerStarted","Data":"b70f3f77617be0e967a5236636534d73425a03e65376c4ab5f00f3d6ae2e19bf"} Dec 06 09:30:02 crc kubenswrapper[4945]: I1206 09:30:02.201194 4945 generic.go:334] "Generic (PLEG): container finished" podID="41e98db9-5682-4905-8ebe-9910226b8a79" containerID="b771114dcfcbcbf1129fe966dc42e3dff5de4cafb3b182ad32b408934ab7748c" exitCode=0 Dec 06 09:30:02 crc kubenswrapper[4945]: I1206 09:30:02.201304 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" event={"ID":"41e98db9-5682-4905-8ebe-9910226b8a79","Type":"ContainerDied","Data":"b771114dcfcbcbf1129fe966dc42e3dff5de4cafb3b182ad32b408934ab7748c"} Dec 06 09:30:03 crc kubenswrapper[4945]: I1206 09:30:03.849974 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:03 crc kubenswrapper[4945]: I1206 09:30:03.983714 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41e98db9-5682-4905-8ebe-9910226b8a79-secret-volume\") pod \"41e98db9-5682-4905-8ebe-9910226b8a79\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " Dec 06 09:30:03 crc kubenswrapper[4945]: I1206 09:30:03.984078 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41e98db9-5682-4905-8ebe-9910226b8a79-config-volume\") pod \"41e98db9-5682-4905-8ebe-9910226b8a79\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " Dec 06 09:30:03 crc kubenswrapper[4945]: I1206 09:30:03.984214 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66z5d\" (UniqueName: \"kubernetes.io/projected/41e98db9-5682-4905-8ebe-9910226b8a79-kube-api-access-66z5d\") pod \"41e98db9-5682-4905-8ebe-9910226b8a79\" (UID: \"41e98db9-5682-4905-8ebe-9910226b8a79\") " Dec 06 09:30:03 crc kubenswrapper[4945]: I1206 09:30:03.984921 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41e98db9-5682-4905-8ebe-9910226b8a79-config-volume" (OuterVolumeSpecName: "config-volume") pod "41e98db9-5682-4905-8ebe-9910226b8a79" (UID: "41e98db9-5682-4905-8ebe-9910226b8a79"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 09:30:03 crc kubenswrapper[4945]: I1206 09:30:03.989929 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41e98db9-5682-4905-8ebe-9910226b8a79-kube-api-access-66z5d" (OuterVolumeSpecName: "kube-api-access-66z5d") pod "41e98db9-5682-4905-8ebe-9910226b8a79" (UID: "41e98db9-5682-4905-8ebe-9910226b8a79"). InnerVolumeSpecName "kube-api-access-66z5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:30:03 crc kubenswrapper[4945]: I1206 09:30:03.990611 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e98db9-5682-4905-8ebe-9910226b8a79-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "41e98db9-5682-4905-8ebe-9910226b8a79" (UID: "41e98db9-5682-4905-8ebe-9910226b8a79"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:30:04 crc kubenswrapper[4945]: I1206 09:30:04.087098 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41e98db9-5682-4905-8ebe-9910226b8a79-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 09:30:04 crc kubenswrapper[4945]: I1206 09:30:04.087141 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41e98db9-5682-4905-8ebe-9910226b8a79-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 09:30:04 crc kubenswrapper[4945]: I1206 09:30:04.087156 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66z5d\" (UniqueName: \"kubernetes.io/projected/41e98db9-5682-4905-8ebe-9910226b8a79-kube-api-access-66z5d\") on node \"crc\" DevicePath \"\"" Dec 06 09:30:04 crc kubenswrapper[4945]: I1206 09:30:04.266635 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" event={"ID":"41e98db9-5682-4905-8ebe-9910226b8a79","Type":"ContainerDied","Data":"b70f3f77617be0e967a5236636534d73425a03e65376c4ab5f00f3d6ae2e19bf"} Dec 06 09:30:04 crc kubenswrapper[4945]: I1206 09:30:04.266755 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b70f3f77617be0e967a5236636534d73425a03e65376c4ab5f00f3d6ae2e19bf" Dec 06 09:30:04 crc kubenswrapper[4945]: I1206 09:30:04.266944 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416890-mz44w" Dec 06 09:30:04 crc kubenswrapper[4945]: I1206 09:30:04.942443 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl"] Dec 06 09:30:04 crc kubenswrapper[4945]: I1206 09:30:04.968096 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416845-v4rhl"] Dec 06 09:30:06 crc kubenswrapper[4945]: I1206 09:30:06.967137 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d680a703-dd11-479b-bd30-c2759e64d8cd" path="/var/lib/kubelet/pods/d680a703-dd11-479b-bd30-c2759e64d8cd/volumes" Dec 06 09:30:32 crc kubenswrapper[4945]: I1206 09:30:32.226272 4945 scope.go:117] "RemoveContainer" containerID="ab876645010d526220eee00efec543a44be7649ec2c05cac429ea46f1695973e" Dec 06 09:32:08 crc kubenswrapper[4945]: I1206 09:32:08.796269 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:32:08 crc kubenswrapper[4945]: I1206 09:32:08.796883 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.588695 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r95qf"] Dec 06 09:32:17 crc kubenswrapper[4945]: E1206 09:32:17.589625 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41e98db9-5682-4905-8ebe-9910226b8a79" containerName="collect-profiles" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.589638 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="41e98db9-5682-4905-8ebe-9910226b8a79" containerName="collect-profiles" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.589829 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="41e98db9-5682-4905-8ebe-9910226b8a79" containerName="collect-profiles" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.594736 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.600739 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r95qf"] Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.721470 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-catalog-content\") pod \"certified-operators-r95qf\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.721687 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-utilities\") pod \"certified-operators-r95qf\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.721770 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg579\" (UniqueName: \"kubernetes.io/projected/3d263a04-6815-4f2a-8a88-6381e9dc3df2-kube-api-access-rg579\") pod \"certified-operators-r95qf\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.823505 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-utilities\") pod \"certified-operators-r95qf\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.823557 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg579\" (UniqueName: \"kubernetes.io/projected/3d263a04-6815-4f2a-8a88-6381e9dc3df2-kube-api-access-rg579\") pod \"certified-operators-r95qf\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.823672 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-catalog-content\") pod \"certified-operators-r95qf\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.824234 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-catalog-content\") pod \"certified-operators-r95qf\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.824234 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-utilities\") pod \"certified-operators-r95qf\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.847182 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg579\" (UniqueName: \"kubernetes.io/projected/3d263a04-6815-4f2a-8a88-6381e9dc3df2-kube-api-access-rg579\") pod \"certified-operators-r95qf\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:17 crc kubenswrapper[4945]: I1206 09:32:17.956540 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:18 crc kubenswrapper[4945]: I1206 09:32:18.534964 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r95qf"] Dec 06 09:32:19 crc kubenswrapper[4945]: I1206 09:32:19.231802 4945 generic.go:334] "Generic (PLEG): container finished" podID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerID="f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d" exitCode=0 Dec 06 09:32:19 crc kubenswrapper[4945]: I1206 09:32:19.231937 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r95qf" event={"ID":"3d263a04-6815-4f2a-8a88-6381e9dc3df2","Type":"ContainerDied","Data":"f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d"} Dec 06 09:32:19 crc kubenswrapper[4945]: I1206 09:32:19.232086 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r95qf" event={"ID":"3d263a04-6815-4f2a-8a88-6381e9dc3df2","Type":"ContainerStarted","Data":"505508b4a6842ae61b45b6e81c1b1fc2aaf0ef7fb4ee1c86954a7ac5efc412cf"} Dec 06 09:32:19 crc kubenswrapper[4945]: I1206 09:32:19.234030 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 09:32:20 crc kubenswrapper[4945]: I1206 09:32:20.244013 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r95qf" event={"ID":"3d263a04-6815-4f2a-8a88-6381e9dc3df2","Type":"ContainerStarted","Data":"faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115"} Dec 06 09:32:21 crc kubenswrapper[4945]: I1206 09:32:21.253541 4945 generic.go:334] "Generic (PLEG): container finished" podID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerID="faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115" exitCode=0 Dec 06 09:32:21 crc kubenswrapper[4945]: I1206 09:32:21.253610 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r95qf" event={"ID":"3d263a04-6815-4f2a-8a88-6381e9dc3df2","Type":"ContainerDied","Data":"faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115"} Dec 06 09:32:22 crc kubenswrapper[4945]: I1206 09:32:22.265054 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r95qf" event={"ID":"3d263a04-6815-4f2a-8a88-6381e9dc3df2","Type":"ContainerStarted","Data":"49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d"} Dec 06 09:32:22 crc kubenswrapper[4945]: I1206 09:32:22.285190 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r95qf" podStartSLOduration=2.885446615 podStartE2EDuration="5.285172576s" podCreationTimestamp="2025-12-06 09:32:17 +0000 UTC" firstStartedPulling="2025-12-06 09:32:19.233726635 +0000 UTC m=+9552.688587679" lastFinishedPulling="2025-12-06 09:32:21.633452596 +0000 UTC m=+9555.088313640" observedRunningTime="2025-12-06 09:32:22.28236136 +0000 UTC m=+9555.737222414" watchObservedRunningTime="2025-12-06 09:32:22.285172576 +0000 UTC m=+9555.740033610" Dec 06 09:32:27 crc kubenswrapper[4945]: I1206 09:32:27.957635 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:27 crc kubenswrapper[4945]: I1206 09:32:27.958738 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:28 crc kubenswrapper[4945]: I1206 09:32:28.007529 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:28 crc kubenswrapper[4945]: I1206 09:32:28.386895 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:31 crc kubenswrapper[4945]: I1206 09:32:31.590831 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r95qf"] Dec 06 09:32:31 crc kubenswrapper[4945]: I1206 09:32:31.592484 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r95qf" podUID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerName="registry-server" containerID="cri-o://49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d" gracePeriod=2 Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.328459 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.368640 4945 generic.go:334] "Generic (PLEG): container finished" podID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerID="49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d" exitCode=0 Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.368691 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r95qf" event={"ID":"3d263a04-6815-4f2a-8a88-6381e9dc3df2","Type":"ContainerDied","Data":"49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d"} Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.368724 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r95qf" event={"ID":"3d263a04-6815-4f2a-8a88-6381e9dc3df2","Type":"ContainerDied","Data":"505508b4a6842ae61b45b6e81c1b1fc2aaf0ef7fb4ee1c86954a7ac5efc412cf"} Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.368744 4945 scope.go:117] "RemoveContainer" containerID="49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.368757 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r95qf" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.412598 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg579\" (UniqueName: \"kubernetes.io/projected/3d263a04-6815-4f2a-8a88-6381e9dc3df2-kube-api-access-rg579\") pod \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.412928 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-catalog-content\") pod \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.413069 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-utilities\") pod \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\" (UID: \"3d263a04-6815-4f2a-8a88-6381e9dc3df2\") " Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.413898 4945 scope.go:117] "RemoveContainer" containerID="faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.414122 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-utilities" (OuterVolumeSpecName: "utilities") pod "3d263a04-6815-4f2a-8a88-6381e9dc3df2" (UID: "3d263a04-6815-4f2a-8a88-6381e9dc3df2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.424039 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d263a04-6815-4f2a-8a88-6381e9dc3df2-kube-api-access-rg579" (OuterVolumeSpecName: "kube-api-access-rg579") pod "3d263a04-6815-4f2a-8a88-6381e9dc3df2" (UID: "3d263a04-6815-4f2a-8a88-6381e9dc3df2"). InnerVolumeSpecName "kube-api-access-rg579". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.466060 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d263a04-6815-4f2a-8a88-6381e9dc3df2" (UID: "3d263a04-6815-4f2a-8a88-6381e9dc3df2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.475201 4945 scope.go:117] "RemoveContainer" containerID="f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.515429 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg579\" (UniqueName: \"kubernetes.io/projected/3d263a04-6815-4f2a-8a88-6381e9dc3df2-kube-api-access-rg579\") on node \"crc\" DevicePath \"\"" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.515526 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.515546 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d263a04-6815-4f2a-8a88-6381e9dc3df2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.527871 4945 scope.go:117] "RemoveContainer" containerID="49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d" Dec 06 09:32:32 crc kubenswrapper[4945]: E1206 09:32:32.528309 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d\": container with ID starting with 49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d not found: ID does not exist" containerID="49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.528338 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d"} err="failed to get container status \"49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d\": rpc error: code = NotFound desc = could not find container \"49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d\": container with ID starting with 49013a20bdaf7c414933f7a8757306e45acc87c9540a4bcc5d17cb9ccf30612d not found: ID does not exist" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.528357 4945 scope.go:117] "RemoveContainer" containerID="faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115" Dec 06 09:32:32 crc kubenswrapper[4945]: E1206 09:32:32.528910 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115\": container with ID starting with faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115 not found: ID does not exist" containerID="faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.528938 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115"} err="failed to get container status \"faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115\": rpc error: code = NotFound desc = could not find container \"faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115\": container with ID starting with faaaec6a6e2d33831f3cc0e0c403050ccab4ba1801143293c94981f5b0224115 not found: ID does not exist" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.528952 4945 scope.go:117] "RemoveContainer" containerID="f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d" Dec 06 09:32:32 crc kubenswrapper[4945]: E1206 09:32:32.530750 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d\": container with ID starting with f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d not found: ID does not exist" containerID="f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.530813 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d"} err="failed to get container status \"f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d\": rpc error: code = NotFound desc = could not find container \"f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d\": container with ID starting with f9e3a2c422e34527ddc80cf1cd65032b9a9df0764cdc4a7a9c780435e244d39d not found: ID does not exist" Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.711991 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r95qf"] Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.732817 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r95qf"] Dec 06 09:32:32 crc kubenswrapper[4945]: I1206 09:32:32.965415 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" path="/var/lib/kubelet/pods/3d263a04-6815-4f2a-8a88-6381e9dc3df2/volumes" Dec 06 09:32:38 crc kubenswrapper[4945]: I1206 09:32:38.796159 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:32:38 crc kubenswrapper[4945]: I1206 09:32:38.796746 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:33:08 crc kubenswrapper[4945]: I1206 09:33:08.796084 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:33:08 crc kubenswrapper[4945]: I1206 09:33:08.797105 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:33:08 crc kubenswrapper[4945]: I1206 09:33:08.797180 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 09:33:08 crc kubenswrapper[4945]: I1206 09:33:08.798527 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 09:33:08 crc kubenswrapper[4945]: I1206 09:33:08.798609 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" gracePeriod=600 Dec 06 09:33:08 crc kubenswrapper[4945]: E1206 09:33:08.921710 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:33:08 crc kubenswrapper[4945]: I1206 09:33:08.951259 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" exitCode=0 Dec 06 09:33:08 crc kubenswrapper[4945]: I1206 09:33:08.951365 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f"} Dec 06 09:33:08 crc kubenswrapper[4945]: I1206 09:33:08.951409 4945 scope.go:117] "RemoveContainer" containerID="46722cb4b2f771a8c019f69da4a13a14521d53249d80f3919d51feb2c76b849a" Dec 06 09:33:08 crc kubenswrapper[4945]: I1206 09:33:08.952572 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:33:08 crc kubenswrapper[4945]: E1206 09:33:08.953012 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:33:22 crc kubenswrapper[4945]: I1206 09:33:22.954794 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:33:22 crc kubenswrapper[4945]: E1206 09:33:22.956109 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:33:34 crc kubenswrapper[4945]: I1206 09:33:34.953520 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:33:34 crc kubenswrapper[4945]: E1206 09:33:34.954319 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:33:49 crc kubenswrapper[4945]: I1206 09:33:49.953388 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:33:49 crc kubenswrapper[4945]: E1206 09:33:49.954347 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:34:01 crc kubenswrapper[4945]: I1206 09:34:01.953996 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:34:01 crc kubenswrapper[4945]: E1206 09:34:01.954871 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:34:16 crc kubenswrapper[4945]: I1206 09:34:16.960372 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:34:16 crc kubenswrapper[4945]: E1206 09:34:16.962092 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:34:27 crc kubenswrapper[4945]: I1206 09:34:27.953213 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:34:27 crc kubenswrapper[4945]: E1206 09:34:27.953962 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:34:40 crc kubenswrapper[4945]: I1206 09:34:40.954003 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:34:40 crc kubenswrapper[4945]: E1206 09:34:40.955042 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:34:45 crc kubenswrapper[4945]: I1206 09:34:45.849552 4945 generic.go:334] "Generic (PLEG): container finished" podID="1000996d-89a3-43f6-bffc-03fd64593568" containerID="141ee0777a89d8c08f7ff6884c81911edefb7e920eacd6752a41565cdb06fbe5" exitCode=0 Dec 06 09:34:45 crc kubenswrapper[4945]: I1206 09:34:45.849639 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1000996d-89a3-43f6-bffc-03fd64593568","Type":"ContainerDied","Data":"141ee0777a89d8c08f7ff6884c81911edefb7e920eacd6752a41565cdb06fbe5"} Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.385890 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.466992 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ssh-key\") pod \"1000996d-89a3-43f6-bffc-03fd64593568\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.467334 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"1000996d-89a3-43f6-bffc-03fd64593568\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.467484 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-config-data\") pod \"1000996d-89a3-43f6-bffc-03fd64593568\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.467605 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhsjq\" (UniqueName: \"kubernetes.io/projected/1000996d-89a3-43f6-bffc-03fd64593568-kube-api-access-lhsjq\") pod \"1000996d-89a3-43f6-bffc-03fd64593568\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.467690 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-temporary\") pod \"1000996d-89a3-43f6-bffc-03fd64593568\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.467823 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config\") pod \"1000996d-89a3-43f6-bffc-03fd64593568\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.467902 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ca-certs\") pod \"1000996d-89a3-43f6-bffc-03fd64593568\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.467990 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-workdir\") pod \"1000996d-89a3-43f6-bffc-03fd64593568\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.468101 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config-secret\") pod \"1000996d-89a3-43f6-bffc-03fd64593568\" (UID: \"1000996d-89a3-43f6-bffc-03fd64593568\") " Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.468261 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "1000996d-89a3-43f6-bffc-03fd64593568" (UID: "1000996d-89a3-43f6-bffc-03fd64593568"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.468415 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-config-data" (OuterVolumeSpecName: "config-data") pod "1000996d-89a3-43f6-bffc-03fd64593568" (UID: "1000996d-89a3-43f6-bffc-03fd64593568"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.468968 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.469063 4945 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.472541 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "1000996d-89a3-43f6-bffc-03fd64593568" (UID: "1000996d-89a3-43f6-bffc-03fd64593568"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.473657 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1000996d-89a3-43f6-bffc-03fd64593568-kube-api-access-lhsjq" (OuterVolumeSpecName: "kube-api-access-lhsjq") pod "1000996d-89a3-43f6-bffc-03fd64593568" (UID: "1000996d-89a3-43f6-bffc-03fd64593568"). InnerVolumeSpecName "kube-api-access-lhsjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.476430 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "1000996d-89a3-43f6-bffc-03fd64593568" (UID: "1000996d-89a3-43f6-bffc-03fd64593568"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.495324 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1000996d-89a3-43f6-bffc-03fd64593568" (UID: "1000996d-89a3-43f6-bffc-03fd64593568"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.497947 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "1000996d-89a3-43f6-bffc-03fd64593568" (UID: "1000996d-89a3-43f6-bffc-03fd64593568"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.501098 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1000996d-89a3-43f6-bffc-03fd64593568" (UID: "1000996d-89a3-43f6-bffc-03fd64593568"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.519005 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1000996d-89a3-43f6-bffc-03fd64593568" (UID: "1000996d-89a3-43f6-bffc-03fd64593568"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.573607 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.573643 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhsjq\" (UniqueName: \"kubernetes.io/projected/1000996d-89a3-43f6-bffc-03fd64593568-kube-api-access-lhsjq\") on node \"crc\" DevicePath \"\"" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.573658 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.573669 4945 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.573681 4945 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1000996d-89a3-43f6-bffc-03fd64593568-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.573692 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.573704 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1000996d-89a3-43f6-bffc-03fd64593568-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.595577 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.676415 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.873699 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1000996d-89a3-43f6-bffc-03fd64593568","Type":"ContainerDied","Data":"785467adddf6845ec72e009f988d0858cf35b599dc02309bcabe9976ef4c8e68"} Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.873749 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="785467adddf6845ec72e009f988d0858cf35b599dc02309bcabe9976ef4c8e68" Dec 06 09:34:47 crc kubenswrapper[4945]: I1206 09:34:47.873773 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.233081 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 09:34:50 crc kubenswrapper[4945]: E1206 09:34:50.233896 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerName="extract-utilities" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.233914 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerName="extract-utilities" Dec 06 09:34:50 crc kubenswrapper[4945]: E1206 09:34:50.233952 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerName="extract-content" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.233958 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerName="extract-content" Dec 06 09:34:50 crc kubenswrapper[4945]: E1206 09:34:50.233974 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerName="registry-server" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.233980 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerName="registry-server" Dec 06 09:34:50 crc kubenswrapper[4945]: E1206 09:34:50.233993 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1000996d-89a3-43f6-bffc-03fd64593568" containerName="tempest-tests-tempest-tests-runner" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.233998 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1000996d-89a3-43f6-bffc-03fd64593568" containerName="tempest-tests-tempest-tests-runner" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.234189 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d263a04-6815-4f2a-8a88-6381e9dc3df2" containerName="registry-server" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.234208 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1000996d-89a3-43f6-bffc-03fd64593568" containerName="tempest-tests-tempest-tests-runner" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.235126 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.236803 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gxzcx" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.245129 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.335903 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ba384fba-e2b7-4e9b-b16b-28673c1567f9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.336214 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxrdt\" (UniqueName: \"kubernetes.io/projected/ba384fba-e2b7-4e9b-b16b-28673c1567f9-kube-api-access-fxrdt\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ba384fba-e2b7-4e9b-b16b-28673c1567f9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.437951 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxrdt\" (UniqueName: \"kubernetes.io/projected/ba384fba-e2b7-4e9b-b16b-28673c1567f9-kube-api-access-fxrdt\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ba384fba-e2b7-4e9b-b16b-28673c1567f9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.438400 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ba384fba-e2b7-4e9b-b16b-28673c1567f9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.438809 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ba384fba-e2b7-4e9b-b16b-28673c1567f9\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.468803 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxrdt\" (UniqueName: \"kubernetes.io/projected/ba384fba-e2b7-4e9b-b16b-28673c1567f9-kube-api-access-fxrdt\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ba384fba-e2b7-4e9b-b16b-28673c1567f9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.492340 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ba384fba-e2b7-4e9b-b16b-28673c1567f9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 09:34:50 crc kubenswrapper[4945]: I1206 09:34:50.565367 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 09:34:51 crc kubenswrapper[4945]: I1206 09:34:51.096914 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 09:34:51 crc kubenswrapper[4945]: W1206 09:34:51.105215 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba384fba_e2b7_4e9b_b16b_28673c1567f9.slice/crio-a9d61a46af6cbe6bc82df18183b7eb5f59144cd8cc00aff45d7e98af21202453 WatchSource:0}: Error finding container a9d61a46af6cbe6bc82df18183b7eb5f59144cd8cc00aff45d7e98af21202453: Status 404 returned error can't find the container with id a9d61a46af6cbe6bc82df18183b7eb5f59144cd8cc00aff45d7e98af21202453 Dec 06 09:34:51 crc kubenswrapper[4945]: I1206 09:34:51.912564 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"ba384fba-e2b7-4e9b-b16b-28673c1567f9","Type":"ContainerStarted","Data":"a9d61a46af6cbe6bc82df18183b7eb5f59144cd8cc00aff45d7e98af21202453"} Dec 06 09:34:52 crc kubenswrapper[4945]: I1206 09:34:52.923665 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"ba384fba-e2b7-4e9b-b16b-28673c1567f9","Type":"ContainerStarted","Data":"5d0fa8eaa8265dbf831d513f251ba4d359fec5a445aea1833669db8a2ca16007"} Dec 06 09:34:52 crc kubenswrapper[4945]: I1206 09:34:52.938751 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.5074828820000001 podStartE2EDuration="2.938735097s" podCreationTimestamp="2025-12-06 09:34:50 +0000 UTC" firstStartedPulling="2025-12-06 09:34:51.112171475 +0000 UTC m=+9704.567032539" lastFinishedPulling="2025-12-06 09:34:52.54342371 +0000 UTC m=+9705.998284754" observedRunningTime="2025-12-06 09:34:52.93548847 +0000 UTC m=+9706.390349524" watchObservedRunningTime="2025-12-06 09:34:52.938735097 +0000 UTC m=+9706.393596141" Dec 06 09:34:53 crc kubenswrapper[4945]: I1206 09:34:53.954071 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:34:53 crc kubenswrapper[4945]: E1206 09:34:53.954601 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:35:05 crc kubenswrapper[4945]: I1206 09:35:05.953781 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:35:05 crc kubenswrapper[4945]: E1206 09:35:05.954489 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:35:19 crc kubenswrapper[4945]: I1206 09:35:19.954212 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:35:19 crc kubenswrapper[4945]: E1206 09:35:19.956208 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:35:31 crc kubenswrapper[4945]: I1206 09:35:31.954407 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:35:31 crc kubenswrapper[4945]: E1206 09:35:31.956339 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:35:44 crc kubenswrapper[4945]: I1206 09:35:44.953899 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:35:44 crc kubenswrapper[4945]: E1206 09:35:44.954569 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:35:53 crc kubenswrapper[4945]: I1206 09:35:53.948884 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rz6dz/must-gather-5r6s7"] Dec 06 09:35:53 crc kubenswrapper[4945]: I1206 09:35:53.951418 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/must-gather-5r6s7" Dec 06 09:35:53 crc kubenswrapper[4945]: I1206 09:35:53.953707 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rz6dz"/"openshift-service-ca.crt" Dec 06 09:35:53 crc kubenswrapper[4945]: I1206 09:35:53.954604 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rz6dz"/"kube-root-ca.crt" Dec 06 09:35:53 crc kubenswrapper[4945]: I1206 09:35:53.955030 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rz6dz"/"default-dockercfg-gg28t" Dec 06 09:35:53 crc kubenswrapper[4945]: I1206 09:35:53.982616 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rz6dz/must-gather-5r6s7"] Dec 06 09:35:54 crc kubenswrapper[4945]: I1206 09:35:54.058669 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-must-gather-output\") pod \"must-gather-5r6s7\" (UID: \"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1\") " pod="openshift-must-gather-rz6dz/must-gather-5r6s7" Dec 06 09:35:54 crc kubenswrapper[4945]: I1206 09:35:54.058857 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctg7s\" (UniqueName: \"kubernetes.io/projected/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-kube-api-access-ctg7s\") pod \"must-gather-5r6s7\" (UID: \"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1\") " pod="openshift-must-gather-rz6dz/must-gather-5r6s7" Dec 06 09:35:54 crc kubenswrapper[4945]: I1206 09:35:54.160167 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctg7s\" (UniqueName: \"kubernetes.io/projected/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-kube-api-access-ctg7s\") pod \"must-gather-5r6s7\" (UID: \"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1\") " pod="openshift-must-gather-rz6dz/must-gather-5r6s7" Dec 06 09:35:54 crc kubenswrapper[4945]: I1206 09:35:54.160262 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-must-gather-output\") pod \"must-gather-5r6s7\" (UID: \"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1\") " pod="openshift-must-gather-rz6dz/must-gather-5r6s7" Dec 06 09:35:54 crc kubenswrapper[4945]: I1206 09:35:54.161431 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-must-gather-output\") pod \"must-gather-5r6s7\" (UID: \"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1\") " pod="openshift-must-gather-rz6dz/must-gather-5r6s7" Dec 06 09:35:54 crc kubenswrapper[4945]: I1206 09:35:54.182070 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctg7s\" (UniqueName: \"kubernetes.io/projected/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-kube-api-access-ctg7s\") pod \"must-gather-5r6s7\" (UID: \"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1\") " pod="openshift-must-gather-rz6dz/must-gather-5r6s7" Dec 06 09:35:54 crc kubenswrapper[4945]: I1206 09:35:54.271537 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/must-gather-5r6s7" Dec 06 09:35:54 crc kubenswrapper[4945]: I1206 09:35:54.787409 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rz6dz/must-gather-5r6s7"] Dec 06 09:35:55 crc kubenswrapper[4945]: I1206 09:35:55.593847 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rz6dz/must-gather-5r6s7" event={"ID":"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1","Type":"ContainerStarted","Data":"6357c61f12dcf80c52115d18f164011cdf2699fd31648b04f3f7657303ac64b4"} Dec 06 09:35:56 crc kubenswrapper[4945]: I1206 09:35:56.965177 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:35:56 crc kubenswrapper[4945]: E1206 09:35:56.965913 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:35:58 crc kubenswrapper[4945]: I1206 09:35:58.627345 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rz6dz/must-gather-5r6s7" event={"ID":"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1","Type":"ContainerStarted","Data":"c531f85676251f5ba45ec49ca3a92bbbb93c4c8fe11e8535703806fed9954f4a"} Dec 06 09:35:59 crc kubenswrapper[4945]: I1206 09:35:59.638133 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rz6dz/must-gather-5r6s7" event={"ID":"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1","Type":"ContainerStarted","Data":"ff36ec5138ea04fe9a79d48e0fa9d78cc51a47f3146157fba2a8023378ebf6ef"} Dec 06 09:35:59 crc kubenswrapper[4945]: I1206 09:35:59.661389 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rz6dz/must-gather-5r6s7" podStartSLOduration=3.189547546 podStartE2EDuration="6.661369533s" podCreationTimestamp="2025-12-06 09:35:53 +0000 UTC" firstStartedPulling="2025-12-06 09:35:54.798298451 +0000 UTC m=+9768.253159495" lastFinishedPulling="2025-12-06 09:35:58.270120438 +0000 UTC m=+9771.724981482" observedRunningTime="2025-12-06 09:35:59.651723075 +0000 UTC m=+9773.106584139" watchObservedRunningTime="2025-12-06 09:35:59.661369533 +0000 UTC m=+9773.116230577" Dec 06 09:36:02 crc kubenswrapper[4945]: E1206 09:36:02.211947 4945 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.220:56654->38.129.56.220:39753: write tcp 38.129.56.220:56654->38.129.56.220:39753: write: broken pipe Dec 06 09:36:03 crc kubenswrapper[4945]: I1206 09:36:03.306683 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rz6dz/crc-debug-pxk2b"] Dec 06 09:36:03 crc kubenswrapper[4945]: I1206 09:36:03.309773 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" Dec 06 09:36:03 crc kubenswrapper[4945]: I1206 09:36:03.437716 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnslz\" (UniqueName: \"kubernetes.io/projected/09ab5ade-6453-4366-a6fe-b131df6b5e97-kube-api-access-lnslz\") pod \"crc-debug-pxk2b\" (UID: \"09ab5ade-6453-4366-a6fe-b131df6b5e97\") " pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" Dec 06 09:36:03 crc kubenswrapper[4945]: I1206 09:36:03.438108 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09ab5ade-6453-4366-a6fe-b131df6b5e97-host\") pod \"crc-debug-pxk2b\" (UID: \"09ab5ade-6453-4366-a6fe-b131df6b5e97\") " pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" Dec 06 09:36:03 crc kubenswrapper[4945]: I1206 09:36:03.540298 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09ab5ade-6453-4366-a6fe-b131df6b5e97-host\") pod \"crc-debug-pxk2b\" (UID: \"09ab5ade-6453-4366-a6fe-b131df6b5e97\") " pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" Dec 06 09:36:03 crc kubenswrapper[4945]: I1206 09:36:03.540512 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnslz\" (UniqueName: \"kubernetes.io/projected/09ab5ade-6453-4366-a6fe-b131df6b5e97-kube-api-access-lnslz\") pod \"crc-debug-pxk2b\" (UID: \"09ab5ade-6453-4366-a6fe-b131df6b5e97\") " pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" Dec 06 09:36:03 crc kubenswrapper[4945]: I1206 09:36:03.541093 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09ab5ade-6453-4366-a6fe-b131df6b5e97-host\") pod \"crc-debug-pxk2b\" (UID: \"09ab5ade-6453-4366-a6fe-b131df6b5e97\") " pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" Dec 06 09:36:03 crc kubenswrapper[4945]: I1206 09:36:03.564981 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnslz\" (UniqueName: \"kubernetes.io/projected/09ab5ade-6453-4366-a6fe-b131df6b5e97-kube-api-access-lnslz\") pod \"crc-debug-pxk2b\" (UID: \"09ab5ade-6453-4366-a6fe-b131df6b5e97\") " pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" Dec 06 09:36:03 crc kubenswrapper[4945]: I1206 09:36:03.631037 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" Dec 06 09:36:04 crc kubenswrapper[4945]: I1206 09:36:04.691183 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" event={"ID":"09ab5ade-6453-4366-a6fe-b131df6b5e97","Type":"ContainerStarted","Data":"00d60b577653b524039a095fa44fb44433a0692cdedc064acba2501c4cc61b76"} Dec 06 09:36:10 crc kubenswrapper[4945]: I1206 09:36:10.956728 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:36:10 crc kubenswrapper[4945]: E1206 09:36:10.957557 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:36:14 crc kubenswrapper[4945]: I1206 09:36:14.807149 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" event={"ID":"09ab5ade-6453-4366-a6fe-b131df6b5e97","Type":"ContainerStarted","Data":"0c4dafdb4e99fef8763676a8f5b0b180a2ecccc8a12750c9da3f31686cf532cf"} Dec 06 09:36:14 crc kubenswrapper[4945]: I1206 09:36:14.822632 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" podStartSLOduration=1.213094574 podStartE2EDuration="11.82261359s" podCreationTimestamp="2025-12-06 09:36:03 +0000 UTC" firstStartedPulling="2025-12-06 09:36:03.674022315 +0000 UTC m=+9777.128883359" lastFinishedPulling="2025-12-06 09:36:14.283541331 +0000 UTC m=+9787.738402375" observedRunningTime="2025-12-06 09:36:14.819038384 +0000 UTC m=+9788.273899448" watchObservedRunningTime="2025-12-06 09:36:14.82261359 +0000 UTC m=+9788.277474634" Dec 06 09:36:23 crc kubenswrapper[4945]: I1206 09:36:23.954266 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:36:23 crc kubenswrapper[4945]: E1206 09:36:23.954986 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:36:34 crc kubenswrapper[4945]: I1206 09:36:34.954155 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:36:34 crc kubenswrapper[4945]: E1206 09:36:34.955140 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:36:36 crc kubenswrapper[4945]: I1206 09:36:36.001979 4945 generic.go:334] "Generic (PLEG): container finished" podID="09ab5ade-6453-4366-a6fe-b131df6b5e97" containerID="0c4dafdb4e99fef8763676a8f5b0b180a2ecccc8a12750c9da3f31686cf532cf" exitCode=0 Dec 06 09:36:36 crc kubenswrapper[4945]: I1206 09:36:36.002082 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" event={"ID":"09ab5ade-6453-4366-a6fe-b131df6b5e97","Type":"ContainerDied","Data":"0c4dafdb4e99fef8763676a8f5b0b180a2ecccc8a12750c9da3f31686cf532cf"} Dec 06 09:36:37 crc kubenswrapper[4945]: I1206 09:36:37.144610 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" Dec 06 09:36:37 crc kubenswrapper[4945]: I1206 09:36:37.190990 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rz6dz/crc-debug-pxk2b"] Dec 06 09:36:37 crc kubenswrapper[4945]: I1206 09:36:37.202594 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rz6dz/crc-debug-pxk2b"] Dec 06 09:36:37 crc kubenswrapper[4945]: I1206 09:36:37.246019 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09ab5ade-6453-4366-a6fe-b131df6b5e97-host\") pod \"09ab5ade-6453-4366-a6fe-b131df6b5e97\" (UID: \"09ab5ade-6453-4366-a6fe-b131df6b5e97\") " Dec 06 09:36:37 crc kubenswrapper[4945]: I1206 09:36:37.246124 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnslz\" (UniqueName: \"kubernetes.io/projected/09ab5ade-6453-4366-a6fe-b131df6b5e97-kube-api-access-lnslz\") pod \"09ab5ade-6453-4366-a6fe-b131df6b5e97\" (UID: \"09ab5ade-6453-4366-a6fe-b131df6b5e97\") " Dec 06 09:36:37 crc kubenswrapper[4945]: I1206 09:36:37.247314 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09ab5ade-6453-4366-a6fe-b131df6b5e97-host" (OuterVolumeSpecName: "host") pod "09ab5ade-6453-4366-a6fe-b131df6b5e97" (UID: "09ab5ade-6453-4366-a6fe-b131df6b5e97"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 09:36:37 crc kubenswrapper[4945]: I1206 09:36:37.252543 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ab5ade-6453-4366-a6fe-b131df6b5e97-kube-api-access-lnslz" (OuterVolumeSpecName: "kube-api-access-lnslz") pod "09ab5ade-6453-4366-a6fe-b131df6b5e97" (UID: "09ab5ade-6453-4366-a6fe-b131df6b5e97"). InnerVolumeSpecName "kube-api-access-lnslz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:36:37 crc kubenswrapper[4945]: I1206 09:36:37.348833 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09ab5ade-6453-4366-a6fe-b131df6b5e97-host\") on node \"crc\" DevicePath \"\"" Dec 06 09:36:37 crc kubenswrapper[4945]: I1206 09:36:37.348872 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnslz\" (UniqueName: \"kubernetes.io/projected/09ab5ade-6453-4366-a6fe-b131df6b5e97-kube-api-access-lnslz\") on node \"crc\" DevicePath \"\"" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.020928 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00d60b577653b524039a095fa44fb44433a0692cdedc064acba2501c4cc61b76" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.020990 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/crc-debug-pxk2b" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.358746 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rz6dz/crc-debug-kq9c4"] Dec 06 09:36:38 crc kubenswrapper[4945]: E1206 09:36:38.359746 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ab5ade-6453-4366-a6fe-b131df6b5e97" containerName="container-00" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.359762 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ab5ade-6453-4366-a6fe-b131df6b5e97" containerName="container-00" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.359977 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="09ab5ade-6453-4366-a6fe-b131df6b5e97" containerName="container-00" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.360741 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.471299 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/991f5d78-0731-4947-bdf8-835072d61d31-host\") pod \"crc-debug-kq9c4\" (UID: \"991f5d78-0731-4947-bdf8-835072d61d31\") " pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.471662 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9fsf\" (UniqueName: \"kubernetes.io/projected/991f5d78-0731-4947-bdf8-835072d61d31-kube-api-access-m9fsf\") pod \"crc-debug-kq9c4\" (UID: \"991f5d78-0731-4947-bdf8-835072d61d31\") " pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.573511 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/991f5d78-0731-4947-bdf8-835072d61d31-host\") pod \"crc-debug-kq9c4\" (UID: \"991f5d78-0731-4947-bdf8-835072d61d31\") " pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.573666 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/991f5d78-0731-4947-bdf8-835072d61d31-host\") pod \"crc-debug-kq9c4\" (UID: \"991f5d78-0731-4947-bdf8-835072d61d31\") " pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.573933 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9fsf\" (UniqueName: \"kubernetes.io/projected/991f5d78-0731-4947-bdf8-835072d61d31-kube-api-access-m9fsf\") pod \"crc-debug-kq9c4\" (UID: \"991f5d78-0731-4947-bdf8-835072d61d31\") " pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.594074 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9fsf\" (UniqueName: \"kubernetes.io/projected/991f5d78-0731-4947-bdf8-835072d61d31-kube-api-access-m9fsf\") pod \"crc-debug-kq9c4\" (UID: \"991f5d78-0731-4947-bdf8-835072d61d31\") " pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.680562 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" Dec 06 09:36:38 crc kubenswrapper[4945]: I1206 09:36:38.966447 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ab5ade-6453-4366-a6fe-b131df6b5e97" path="/var/lib/kubelet/pods/09ab5ade-6453-4366-a6fe-b131df6b5e97/volumes" Dec 06 09:36:39 crc kubenswrapper[4945]: I1206 09:36:39.031562 4945 generic.go:334] "Generic (PLEG): container finished" podID="991f5d78-0731-4947-bdf8-835072d61d31" containerID="b6208a98c4f4ebdfa89d92d6cae98cf9658c3c1d5bfeefe96334d0242bbb7da2" exitCode=1 Dec 06 09:36:39 crc kubenswrapper[4945]: I1206 09:36:39.031602 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" event={"ID":"991f5d78-0731-4947-bdf8-835072d61d31","Type":"ContainerDied","Data":"b6208a98c4f4ebdfa89d92d6cae98cf9658c3c1d5bfeefe96334d0242bbb7da2"} Dec 06 09:36:39 crc kubenswrapper[4945]: I1206 09:36:39.031629 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" event={"ID":"991f5d78-0731-4947-bdf8-835072d61d31","Type":"ContainerStarted","Data":"d54ab90001d09614697a9e6bc51d825e7d98844e43919eb17f45f73b1dd12962"} Dec 06 09:36:39 crc kubenswrapper[4945]: I1206 09:36:39.076824 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rz6dz/crc-debug-kq9c4"] Dec 06 09:36:39 crc kubenswrapper[4945]: I1206 09:36:39.091682 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rz6dz/crc-debug-kq9c4"] Dec 06 09:36:40 crc kubenswrapper[4945]: I1206 09:36:40.160269 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" Dec 06 09:36:40 crc kubenswrapper[4945]: I1206 09:36:40.206146 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/991f5d78-0731-4947-bdf8-835072d61d31-host\") pod \"991f5d78-0731-4947-bdf8-835072d61d31\" (UID: \"991f5d78-0731-4947-bdf8-835072d61d31\") " Dec 06 09:36:40 crc kubenswrapper[4945]: I1206 09:36:40.206242 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9fsf\" (UniqueName: \"kubernetes.io/projected/991f5d78-0731-4947-bdf8-835072d61d31-kube-api-access-m9fsf\") pod \"991f5d78-0731-4947-bdf8-835072d61d31\" (UID: \"991f5d78-0731-4947-bdf8-835072d61d31\") " Dec 06 09:36:40 crc kubenswrapper[4945]: I1206 09:36:40.206236 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/991f5d78-0731-4947-bdf8-835072d61d31-host" (OuterVolumeSpecName: "host") pod "991f5d78-0731-4947-bdf8-835072d61d31" (UID: "991f5d78-0731-4947-bdf8-835072d61d31"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 09:36:40 crc kubenswrapper[4945]: I1206 09:36:40.206837 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/991f5d78-0731-4947-bdf8-835072d61d31-host\") on node \"crc\" DevicePath \"\"" Dec 06 09:36:40 crc kubenswrapper[4945]: I1206 09:36:40.215512 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/991f5d78-0731-4947-bdf8-835072d61d31-kube-api-access-m9fsf" (OuterVolumeSpecName: "kube-api-access-m9fsf") pod "991f5d78-0731-4947-bdf8-835072d61d31" (UID: "991f5d78-0731-4947-bdf8-835072d61d31"). InnerVolumeSpecName "kube-api-access-m9fsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:36:40 crc kubenswrapper[4945]: I1206 09:36:40.308620 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9fsf\" (UniqueName: \"kubernetes.io/projected/991f5d78-0731-4947-bdf8-835072d61d31-kube-api-access-m9fsf\") on node \"crc\" DevicePath \"\"" Dec 06 09:36:40 crc kubenswrapper[4945]: I1206 09:36:40.967433 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="991f5d78-0731-4947-bdf8-835072d61d31" path="/var/lib/kubelet/pods/991f5d78-0731-4947-bdf8-835072d61d31/volumes" Dec 06 09:36:41 crc kubenswrapper[4945]: I1206 09:36:41.049981 4945 scope.go:117] "RemoveContainer" containerID="b6208a98c4f4ebdfa89d92d6cae98cf9658c3c1d5bfeefe96334d0242bbb7da2" Dec 06 09:36:41 crc kubenswrapper[4945]: I1206 09:36:41.050031 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/crc-debug-kq9c4" Dec 06 09:36:46 crc kubenswrapper[4945]: I1206 09:36:46.962397 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:36:46 crc kubenswrapper[4945]: E1206 09:36:46.963160 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:36:57 crc kubenswrapper[4945]: I1206 09:36:57.738752 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9lfb8"] Dec 06 09:36:57 crc kubenswrapper[4945]: E1206 09:36:57.739723 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="991f5d78-0731-4947-bdf8-835072d61d31" containerName="container-00" Dec 06 09:36:57 crc kubenswrapper[4945]: I1206 09:36:57.739736 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="991f5d78-0731-4947-bdf8-835072d61d31" containerName="container-00" Dec 06 09:36:57 crc kubenswrapper[4945]: I1206 09:36:57.739968 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="991f5d78-0731-4947-bdf8-835072d61d31" containerName="container-00" Dec 06 09:36:57 crc kubenswrapper[4945]: I1206 09:36:57.741639 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:57 crc kubenswrapper[4945]: I1206 09:36:57.773976 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9lfb8"] Dec 06 09:36:57 crc kubenswrapper[4945]: I1206 09:36:57.917780 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-catalog-content\") pod \"community-operators-9lfb8\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:57 crc kubenswrapper[4945]: I1206 09:36:57.917940 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlqv6\" (UniqueName: \"kubernetes.io/projected/7d68d6a4-2dc9-4c88-9af1-160b466cab03-kube-api-access-tlqv6\") pod \"community-operators-9lfb8\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:57 crc kubenswrapper[4945]: I1206 09:36:57.917973 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-utilities\") pod \"community-operators-9lfb8\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:57 crc kubenswrapper[4945]: I1206 09:36:57.952926 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:36:57 crc kubenswrapper[4945]: E1206 09:36:57.953225 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:36:58 crc kubenswrapper[4945]: I1206 09:36:58.020238 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlqv6\" (UniqueName: \"kubernetes.io/projected/7d68d6a4-2dc9-4c88-9af1-160b466cab03-kube-api-access-tlqv6\") pod \"community-operators-9lfb8\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:58 crc kubenswrapper[4945]: I1206 09:36:58.020316 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-utilities\") pod \"community-operators-9lfb8\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:58 crc kubenswrapper[4945]: I1206 09:36:58.020487 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-catalog-content\") pod \"community-operators-9lfb8\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:58 crc kubenswrapper[4945]: I1206 09:36:58.021035 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-utilities\") pod \"community-operators-9lfb8\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:58 crc kubenswrapper[4945]: I1206 09:36:58.021095 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-catalog-content\") pod \"community-operators-9lfb8\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:58 crc kubenswrapper[4945]: I1206 09:36:58.046500 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlqv6\" (UniqueName: \"kubernetes.io/projected/7d68d6a4-2dc9-4c88-9af1-160b466cab03-kube-api-access-tlqv6\") pod \"community-operators-9lfb8\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:58 crc kubenswrapper[4945]: I1206 09:36:58.079657 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:36:58 crc kubenswrapper[4945]: I1206 09:36:58.664051 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9lfb8"] Dec 06 09:36:58 crc kubenswrapper[4945]: W1206 09:36:58.664864 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d68d6a4_2dc9_4c88_9af1_160b466cab03.slice/crio-01dc7394e0d00e6855e4bd1d8b4ebe65a959f5a6a2b330630526dc35e72d5741 WatchSource:0}: Error finding container 01dc7394e0d00e6855e4bd1d8b4ebe65a959f5a6a2b330630526dc35e72d5741: Status 404 returned error can't find the container with id 01dc7394e0d00e6855e4bd1d8b4ebe65a959f5a6a2b330630526dc35e72d5741 Dec 06 09:36:59 crc kubenswrapper[4945]: I1206 09:36:59.218306 4945 generic.go:334] "Generic (PLEG): container finished" podID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerID="24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea" exitCode=0 Dec 06 09:36:59 crc kubenswrapper[4945]: I1206 09:36:59.218371 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lfb8" event={"ID":"7d68d6a4-2dc9-4c88-9af1-160b466cab03","Type":"ContainerDied","Data":"24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea"} Dec 06 09:36:59 crc kubenswrapper[4945]: I1206 09:36:59.218580 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lfb8" event={"ID":"7d68d6a4-2dc9-4c88-9af1-160b466cab03","Type":"ContainerStarted","Data":"01dc7394e0d00e6855e4bd1d8b4ebe65a959f5a6a2b330630526dc35e72d5741"} Dec 06 09:37:00 crc kubenswrapper[4945]: I1206 09:37:00.230009 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lfb8" event={"ID":"7d68d6a4-2dc9-4c88-9af1-160b466cab03","Type":"ContainerStarted","Data":"6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37"} Dec 06 09:37:01 crc kubenswrapper[4945]: I1206 09:37:01.245767 4945 generic.go:334] "Generic (PLEG): container finished" podID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerID="6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37" exitCode=0 Dec 06 09:37:01 crc kubenswrapper[4945]: I1206 09:37:01.245852 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lfb8" event={"ID":"7d68d6a4-2dc9-4c88-9af1-160b466cab03","Type":"ContainerDied","Data":"6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37"} Dec 06 09:37:02 crc kubenswrapper[4945]: I1206 09:37:02.257266 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lfb8" event={"ID":"7d68d6a4-2dc9-4c88-9af1-160b466cab03","Type":"ContainerStarted","Data":"807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a"} Dec 06 09:37:02 crc kubenswrapper[4945]: I1206 09:37:02.275679 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9lfb8" podStartSLOduration=2.859527119 podStartE2EDuration="5.2756592s" podCreationTimestamp="2025-12-06 09:36:57 +0000 UTC" firstStartedPulling="2025-12-06 09:36:59.220675574 +0000 UTC m=+9832.675536608" lastFinishedPulling="2025-12-06 09:37:01.636807645 +0000 UTC m=+9835.091668689" observedRunningTime="2025-12-06 09:37:02.274555761 +0000 UTC m=+9835.729416825" watchObservedRunningTime="2025-12-06 09:37:02.2756592 +0000 UTC m=+9835.730520244" Dec 06 09:37:08 crc kubenswrapper[4945]: I1206 09:37:08.080666 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:37:08 crc kubenswrapper[4945]: I1206 09:37:08.081371 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:37:08 crc kubenswrapper[4945]: I1206 09:37:08.128124 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:37:08 crc kubenswrapper[4945]: I1206 09:37:08.356949 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:37:08 crc kubenswrapper[4945]: I1206 09:37:08.401381 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9lfb8"] Dec 06 09:37:09 crc kubenswrapper[4945]: I1206 09:37:09.953661 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:37:09 crc kubenswrapper[4945]: E1206 09:37:09.954223 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:37:10 crc kubenswrapper[4945]: I1206 09:37:10.328511 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9lfb8" podUID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerName="registry-server" containerID="cri-o://807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a" gracePeriod=2 Dec 06 09:37:10 crc kubenswrapper[4945]: I1206 09:37:10.843574 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:37:10 crc kubenswrapper[4945]: I1206 09:37:10.992634 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-catalog-content\") pod \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " Dec 06 09:37:10 crc kubenswrapper[4945]: I1206 09:37:10.992882 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlqv6\" (UniqueName: \"kubernetes.io/projected/7d68d6a4-2dc9-4c88-9af1-160b466cab03-kube-api-access-tlqv6\") pod \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " Dec 06 09:37:10 crc kubenswrapper[4945]: I1206 09:37:10.992981 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-utilities\") pod \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\" (UID: \"7d68d6a4-2dc9-4c88-9af1-160b466cab03\") " Dec 06 09:37:10 crc kubenswrapper[4945]: I1206 09:37:10.993844 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-utilities" (OuterVolumeSpecName: "utilities") pod "7d68d6a4-2dc9-4c88-9af1-160b466cab03" (UID: "7d68d6a4-2dc9-4c88-9af1-160b466cab03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:37:10 crc kubenswrapper[4945]: I1206 09:37:10.994361 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.001519 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d68d6a4-2dc9-4c88-9af1-160b466cab03-kube-api-access-tlqv6" (OuterVolumeSpecName: "kube-api-access-tlqv6") pod "7d68d6a4-2dc9-4c88-9af1-160b466cab03" (UID: "7d68d6a4-2dc9-4c88-9af1-160b466cab03"). InnerVolumeSpecName "kube-api-access-tlqv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.045221 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d68d6a4-2dc9-4c88-9af1-160b466cab03" (UID: "7d68d6a4-2dc9-4c88-9af1-160b466cab03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.098902 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d68d6a4-2dc9-4c88-9af1-160b466cab03-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.098979 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlqv6\" (UniqueName: \"kubernetes.io/projected/7d68d6a4-2dc9-4c88-9af1-160b466cab03-kube-api-access-tlqv6\") on node \"crc\" DevicePath \"\"" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.348007 4945 generic.go:334] "Generic (PLEG): container finished" podID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerID="807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a" exitCode=0 Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.348059 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lfb8" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.348062 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lfb8" event={"ID":"7d68d6a4-2dc9-4c88-9af1-160b466cab03","Type":"ContainerDied","Data":"807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a"} Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.348117 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lfb8" event={"ID":"7d68d6a4-2dc9-4c88-9af1-160b466cab03","Type":"ContainerDied","Data":"01dc7394e0d00e6855e4bd1d8b4ebe65a959f5a6a2b330630526dc35e72d5741"} Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.348141 4945 scope.go:117] "RemoveContainer" containerID="807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.379375 4945 scope.go:117] "RemoveContainer" containerID="6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.391795 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9lfb8"] Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.399973 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9lfb8"] Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.408570 4945 scope.go:117] "RemoveContainer" containerID="24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.439661 4945 scope.go:117] "RemoveContainer" containerID="807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a" Dec 06 09:37:11 crc kubenswrapper[4945]: E1206 09:37:11.440138 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a\": container with ID starting with 807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a not found: ID does not exist" containerID="807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.440186 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a"} err="failed to get container status \"807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a\": rpc error: code = NotFound desc = could not find container \"807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a\": container with ID starting with 807b62cd8d16dd4d1de46d69207fa064226168d1affcc66feb059d79e2269d0a not found: ID does not exist" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.440213 4945 scope.go:117] "RemoveContainer" containerID="6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37" Dec 06 09:37:11 crc kubenswrapper[4945]: E1206 09:37:11.440582 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37\": container with ID starting with 6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37 not found: ID does not exist" containerID="6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.440618 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37"} err="failed to get container status \"6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37\": rpc error: code = NotFound desc = could not find container \"6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37\": container with ID starting with 6e64d20a9f312ebf46a5267b1c296716fba0fde3e7c198f159b98d0b0d452a37 not found: ID does not exist" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.440638 4945 scope.go:117] "RemoveContainer" containerID="24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea" Dec 06 09:37:11 crc kubenswrapper[4945]: E1206 09:37:11.441064 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea\": container with ID starting with 24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea not found: ID does not exist" containerID="24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea" Dec 06 09:37:11 crc kubenswrapper[4945]: I1206 09:37:11.441087 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea"} err="failed to get container status \"24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea\": rpc error: code = NotFound desc = could not find container \"24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea\": container with ID starting with 24a72bddffd1c73e178b5c9787dcedcdb06a116085219c26bc6102d2ac1456ea not found: ID does not exist" Dec 06 09:37:12 crc kubenswrapper[4945]: I1206 09:37:12.965415 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" path="/var/lib/kubelet/pods/7d68d6a4-2dc9-4c88-9af1-160b466cab03/volumes" Dec 06 09:37:21 crc kubenswrapper[4945]: I1206 09:37:21.953427 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:37:21 crc kubenswrapper[4945]: E1206 09:37:21.955254 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:37:34 crc kubenswrapper[4945]: I1206 09:37:34.953809 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:37:34 crc kubenswrapper[4945]: E1206 09:37:34.954673 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:37:47 crc kubenswrapper[4945]: I1206 09:37:47.953497 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:37:47 crc kubenswrapper[4945]: E1206 09:37:47.954479 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:37:56 crc kubenswrapper[4945]: I1206 09:37:56.816734 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzp6"] Dec 06 09:37:56 crc kubenswrapper[4945]: E1206 09:37:56.819502 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerName="extract-content" Dec 06 09:37:56 crc kubenswrapper[4945]: I1206 09:37:56.819622 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerName="extract-content" Dec 06 09:37:56 crc kubenswrapper[4945]: E1206 09:37:56.819723 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerName="registry-server" Dec 06 09:37:56 crc kubenswrapper[4945]: I1206 09:37:56.819799 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerName="registry-server" Dec 06 09:37:56 crc kubenswrapper[4945]: E1206 09:37:56.819908 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerName="extract-utilities" Dec 06 09:37:56 crc kubenswrapper[4945]: I1206 09:37:56.819995 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerName="extract-utilities" Dec 06 09:37:56 crc kubenswrapper[4945]: I1206 09:37:56.820385 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d68d6a4-2dc9-4c88-9af1-160b466cab03" containerName="registry-server" Dec 06 09:37:56 crc kubenswrapper[4945]: I1206 09:37:56.822446 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:56 crc kubenswrapper[4945]: I1206 09:37:56.854216 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzp6"] Dec 06 09:37:56 crc kubenswrapper[4945]: I1206 09:37:56.956988 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-utilities\") pod \"redhat-marketplace-bnzp6\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:56 crc kubenswrapper[4945]: I1206 09:37:56.957086 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-catalog-content\") pod \"redhat-marketplace-bnzp6\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:56 crc kubenswrapper[4945]: I1206 09:37:56.957158 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ph7k\" (UniqueName: \"kubernetes.io/projected/ddd9f26a-139e-476b-a850-258ec8da775b-kube-api-access-8ph7k\") pod \"redhat-marketplace-bnzp6\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.058862 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-utilities\") pod \"redhat-marketplace-bnzp6\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.058984 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-catalog-content\") pod \"redhat-marketplace-bnzp6\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.059054 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ph7k\" (UniqueName: \"kubernetes.io/projected/ddd9f26a-139e-476b-a850-258ec8da775b-kube-api-access-8ph7k\") pod \"redhat-marketplace-bnzp6\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.059426 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-utilities\") pod \"redhat-marketplace-bnzp6\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.059759 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-catalog-content\") pod \"redhat-marketplace-bnzp6\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.079254 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ph7k\" (UniqueName: \"kubernetes.io/projected/ddd9f26a-139e-476b-a850-258ec8da775b-kube-api-access-8ph7k\") pod \"redhat-marketplace-bnzp6\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.146863 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.403895 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jvqnh"] Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.406602 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.418723 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jvqnh"] Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.466860 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwnbx\" (UniqueName: \"kubernetes.io/projected/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-kube-api-access-pwnbx\") pod \"redhat-operators-jvqnh\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.466966 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-utilities\") pod \"redhat-operators-jvqnh\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.467024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-catalog-content\") pod \"redhat-operators-jvqnh\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.568716 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-catalog-content\") pod \"redhat-operators-jvqnh\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.568831 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwnbx\" (UniqueName: \"kubernetes.io/projected/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-kube-api-access-pwnbx\") pod \"redhat-operators-jvqnh\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.568921 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-utilities\") pod \"redhat-operators-jvqnh\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.569366 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-catalog-content\") pod \"redhat-operators-jvqnh\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.569431 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-utilities\") pod \"redhat-operators-jvqnh\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.590159 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwnbx\" (UniqueName: \"kubernetes.io/projected/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-kube-api-access-pwnbx\") pod \"redhat-operators-jvqnh\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.674872 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzp6"] Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.731724 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:37:57 crc kubenswrapper[4945]: I1206 09:37:57.866672 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzp6" event={"ID":"ddd9f26a-139e-476b-a850-258ec8da775b","Type":"ContainerStarted","Data":"cf77c2ffc14d53d26fe778f242ae1302e159a8da226b8144961f116edabc0ebb"} Dec 06 09:37:58 crc kubenswrapper[4945]: I1206 09:37:58.257052 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jvqnh"] Dec 06 09:37:58 crc kubenswrapper[4945]: W1206 09:37:58.261202 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb4c1fda_a823_4eb3_9ce9_ee374ec2aefa.slice/crio-4ce1e448dc81e42278d849863594dc0c49e0611d0bd3bf1af26869a18b3c6c74 WatchSource:0}: Error finding container 4ce1e448dc81e42278d849863594dc0c49e0611d0bd3bf1af26869a18b3c6c74: Status 404 returned error can't find the container with id 4ce1e448dc81e42278d849863594dc0c49e0611d0bd3bf1af26869a18b3c6c74 Dec 06 09:37:58 crc kubenswrapper[4945]: I1206 09:37:58.877780 4945 generic.go:334] "Generic (PLEG): container finished" podID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerID="2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4" exitCode=0 Dec 06 09:37:58 crc kubenswrapper[4945]: I1206 09:37:58.877863 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvqnh" event={"ID":"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa","Type":"ContainerDied","Data":"2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4"} Dec 06 09:37:58 crc kubenswrapper[4945]: I1206 09:37:58.877890 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvqnh" event={"ID":"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa","Type":"ContainerStarted","Data":"4ce1e448dc81e42278d849863594dc0c49e0611d0bd3bf1af26869a18b3c6c74"} Dec 06 09:37:58 crc kubenswrapper[4945]: I1206 09:37:58.879942 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 09:37:58 crc kubenswrapper[4945]: I1206 09:37:58.881457 4945 generic.go:334] "Generic (PLEG): container finished" podID="ddd9f26a-139e-476b-a850-258ec8da775b" containerID="2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb" exitCode=0 Dec 06 09:37:58 crc kubenswrapper[4945]: I1206 09:37:58.881498 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzp6" event={"ID":"ddd9f26a-139e-476b-a850-258ec8da775b","Type":"ContainerDied","Data":"2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb"} Dec 06 09:37:59 crc kubenswrapper[4945]: I1206 09:37:59.909647 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzp6" event={"ID":"ddd9f26a-139e-476b-a850-258ec8da775b","Type":"ContainerStarted","Data":"c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef"} Dec 06 09:37:59 crc kubenswrapper[4945]: I1206 09:37:59.913416 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvqnh" event={"ID":"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa","Type":"ContainerStarted","Data":"2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360"} Dec 06 09:38:00 crc kubenswrapper[4945]: I1206 09:38:00.931072 4945 generic.go:334] "Generic (PLEG): container finished" podID="ddd9f26a-139e-476b-a850-258ec8da775b" containerID="c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef" exitCode=0 Dec 06 09:38:00 crc kubenswrapper[4945]: I1206 09:38:00.932614 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzp6" event={"ID":"ddd9f26a-139e-476b-a850-258ec8da775b","Type":"ContainerDied","Data":"c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef"} Dec 06 09:38:00 crc kubenswrapper[4945]: I1206 09:38:00.963206 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:38:00 crc kubenswrapper[4945]: E1206 09:38:00.963673 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:38:01 crc kubenswrapper[4945]: I1206 09:38:01.950210 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzp6" event={"ID":"ddd9f26a-139e-476b-a850-258ec8da775b","Type":"ContainerStarted","Data":"695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91"} Dec 06 09:38:02 crc kubenswrapper[4945]: I1206 09:38:02.002404 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bnzp6" podStartSLOduration=3.567337249 podStartE2EDuration="6.002385214s" podCreationTimestamp="2025-12-06 09:37:56 +0000 UTC" firstStartedPulling="2025-12-06 09:37:58.883042659 +0000 UTC m=+9892.337903703" lastFinishedPulling="2025-12-06 09:38:01.318090624 +0000 UTC m=+9894.772951668" observedRunningTime="2025-12-06 09:38:01.966920737 +0000 UTC m=+9895.421781801" watchObservedRunningTime="2025-12-06 09:38:02.002385214 +0000 UTC m=+9895.457246258" Dec 06 09:38:03 crc kubenswrapper[4945]: I1206 09:38:03.968485 4945 generic.go:334] "Generic (PLEG): container finished" podID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerID="2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360" exitCode=0 Dec 06 09:38:03 crc kubenswrapper[4945]: I1206 09:38:03.968576 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvqnh" event={"ID":"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa","Type":"ContainerDied","Data":"2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360"} Dec 06 09:38:05 crc kubenswrapper[4945]: I1206 09:38:05.993627 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvqnh" event={"ID":"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa","Type":"ContainerStarted","Data":"8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd"} Dec 06 09:38:06 crc kubenswrapper[4945]: I1206 09:38:06.010872 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jvqnh" podStartSLOduration=2.78621056 podStartE2EDuration="9.010854675s" podCreationTimestamp="2025-12-06 09:37:57 +0000 UTC" firstStartedPulling="2025-12-06 09:37:58.87971444 +0000 UTC m=+9892.334575484" lastFinishedPulling="2025-12-06 09:38:05.104358555 +0000 UTC m=+9898.559219599" observedRunningTime="2025-12-06 09:38:06.009035346 +0000 UTC m=+9899.463896390" watchObservedRunningTime="2025-12-06 09:38:06.010854675 +0000 UTC m=+9899.465715719" Dec 06 09:38:07 crc kubenswrapper[4945]: I1206 09:38:07.147044 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:38:07 crc kubenswrapper[4945]: I1206 09:38:07.147346 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:38:07 crc kubenswrapper[4945]: I1206 09:38:07.714832 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:38:07 crc kubenswrapper[4945]: I1206 09:38:07.732407 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:38:07 crc kubenswrapper[4945]: I1206 09:38:07.732549 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:38:08 crc kubenswrapper[4945]: I1206 09:38:08.061194 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:38:08 crc kubenswrapper[4945]: I1206 09:38:08.775042 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jvqnh" podUID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerName="registry-server" probeResult="failure" output=< Dec 06 09:38:08 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Dec 06 09:38:08 crc kubenswrapper[4945]: > Dec 06 09:38:08 crc kubenswrapper[4945]: I1206 09:38:08.795139 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzp6"] Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.029683 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bnzp6" podUID="ddd9f26a-139e-476b-a850-258ec8da775b" containerName="registry-server" containerID="cri-o://695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91" gracePeriod=2 Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.594178 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.678250 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-utilities\") pod \"ddd9f26a-139e-476b-a850-258ec8da775b\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.678370 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ph7k\" (UniqueName: \"kubernetes.io/projected/ddd9f26a-139e-476b-a850-258ec8da775b-kube-api-access-8ph7k\") pod \"ddd9f26a-139e-476b-a850-258ec8da775b\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.678427 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-catalog-content\") pod \"ddd9f26a-139e-476b-a850-258ec8da775b\" (UID: \"ddd9f26a-139e-476b-a850-258ec8da775b\") " Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.682038 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-utilities" (OuterVolumeSpecName: "utilities") pod "ddd9f26a-139e-476b-a850-258ec8da775b" (UID: "ddd9f26a-139e-476b-a850-258ec8da775b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.689331 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddd9f26a-139e-476b-a850-258ec8da775b-kube-api-access-8ph7k" (OuterVolumeSpecName: "kube-api-access-8ph7k") pod "ddd9f26a-139e-476b-a850-258ec8da775b" (UID: "ddd9f26a-139e-476b-a850-258ec8da775b"). InnerVolumeSpecName "kube-api-access-8ph7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.698451 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ddd9f26a-139e-476b-a850-258ec8da775b" (UID: "ddd9f26a-139e-476b-a850-258ec8da775b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.780377 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.780703 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ph7k\" (UniqueName: \"kubernetes.io/projected/ddd9f26a-139e-476b-a850-258ec8da775b-kube-api-access-8ph7k\") on node \"crc\" DevicePath \"\"" Dec 06 09:38:10 crc kubenswrapper[4945]: I1206 09:38:10.780715 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddd9f26a-139e-476b-a850-258ec8da775b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.044169 4945 generic.go:334] "Generic (PLEG): container finished" podID="ddd9f26a-139e-476b-a850-258ec8da775b" containerID="695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91" exitCode=0 Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.044219 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzp6" event={"ID":"ddd9f26a-139e-476b-a850-258ec8da775b","Type":"ContainerDied","Data":"695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91"} Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.044239 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnzp6" Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.044255 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzp6" event={"ID":"ddd9f26a-139e-476b-a850-258ec8da775b","Type":"ContainerDied","Data":"cf77c2ffc14d53d26fe778f242ae1302e159a8da226b8144961f116edabc0ebb"} Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.044291 4945 scope.go:117] "RemoveContainer" containerID="695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91" Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.068442 4945 scope.go:117] "RemoveContainer" containerID="c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef" Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.070098 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzp6"] Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.082443 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzp6"] Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.102207 4945 scope.go:117] "RemoveContainer" containerID="2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb" Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.137674 4945 scope.go:117] "RemoveContainer" containerID="695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91" Dec 06 09:38:11 crc kubenswrapper[4945]: E1206 09:38:11.141406 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91\": container with ID starting with 695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91 not found: ID does not exist" containerID="695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91" Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.141446 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91"} err="failed to get container status \"695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91\": rpc error: code = NotFound desc = could not find container \"695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91\": container with ID starting with 695df80eae93751011469d4476416f642400484ab4f970dd22b309908e47dc91 not found: ID does not exist" Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.141471 4945 scope.go:117] "RemoveContainer" containerID="c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef" Dec 06 09:38:11 crc kubenswrapper[4945]: E1206 09:38:11.141963 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef\": container with ID starting with c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef not found: ID does not exist" containerID="c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef" Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.141994 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef"} err="failed to get container status \"c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef\": rpc error: code = NotFound desc = could not find container \"c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef\": container with ID starting with c57f856c858f0cac62a6527a6c441143be0c9ac5d98c17a4ef4c615df3a3b3ef not found: ID does not exist" Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.142015 4945 scope.go:117] "RemoveContainer" containerID="2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb" Dec 06 09:38:11 crc kubenswrapper[4945]: E1206 09:38:11.142380 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb\": container with ID starting with 2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb not found: ID does not exist" containerID="2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb" Dec 06 09:38:11 crc kubenswrapper[4945]: I1206 09:38:11.142436 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb"} err="failed to get container status \"2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb\": rpc error: code = NotFound desc = could not find container \"2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb\": container with ID starting with 2230354d85eb7df75c990f5aab029694174f501c6ca003bc092f82cce56068bb not found: ID does not exist" Dec 06 09:38:12 crc kubenswrapper[4945]: I1206 09:38:12.966379 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddd9f26a-139e-476b-a850-258ec8da775b" path="/var/lib/kubelet/pods/ddd9f26a-139e-476b-a850-258ec8da775b/volumes" Dec 06 09:38:15 crc kubenswrapper[4945]: I1206 09:38:15.953796 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:38:17 crc kubenswrapper[4945]: I1206 09:38:17.104510 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"64748f113951c547ded8dabe4d2c5762e6623dd8ea101b8b8c6080bfda5b945a"} Dec 06 09:38:18 crc kubenswrapper[4945]: I1206 09:38:18.576346 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:38:18 crc kubenswrapper[4945]: I1206 09:38:18.683172 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:38:18 crc kubenswrapper[4945]: I1206 09:38:18.821490 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jvqnh"] Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.130583 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jvqnh" podUID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerName="registry-server" containerID="cri-o://8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd" gracePeriod=2 Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.725533 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.782164 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwnbx\" (UniqueName: \"kubernetes.io/projected/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-kube-api-access-pwnbx\") pod \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.782447 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-catalog-content\") pod \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.782494 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-utilities\") pod \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\" (UID: \"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa\") " Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.783910 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-utilities" (OuterVolumeSpecName: "utilities") pod "db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" (UID: "db4c1fda-a823-4eb3-9ce9-ee374ec2aefa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.784230 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.789644 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-kube-api-access-pwnbx" (OuterVolumeSpecName: "kube-api-access-pwnbx") pod "db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" (UID: "db4c1fda-a823-4eb3-9ce9-ee374ec2aefa"). InnerVolumeSpecName "kube-api-access-pwnbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.890501 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwnbx\" (UniqueName: \"kubernetes.io/projected/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-kube-api-access-pwnbx\") on node \"crc\" DevicePath \"\"" Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.908547 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" (UID: "db4c1fda-a823-4eb3-9ce9-ee374ec2aefa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:38:20 crc kubenswrapper[4945]: I1206 09:38:20.992400 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.141557 4945 generic.go:334] "Generic (PLEG): container finished" podID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerID="8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd" exitCode=0 Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.141601 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvqnh" event={"ID":"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa","Type":"ContainerDied","Data":"8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd"} Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.141630 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvqnh" event={"ID":"db4c1fda-a823-4eb3-9ce9-ee374ec2aefa","Type":"ContainerDied","Data":"4ce1e448dc81e42278d849863594dc0c49e0611d0bd3bf1af26869a18b3c6c74"} Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.141640 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvqnh" Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.141664 4945 scope.go:117] "RemoveContainer" containerID="8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd" Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.173480 4945 scope.go:117] "RemoveContainer" containerID="2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360" Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.175413 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jvqnh"] Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.186457 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jvqnh"] Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.199175 4945 scope.go:117] "RemoveContainer" containerID="2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4" Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.244728 4945 scope.go:117] "RemoveContainer" containerID="8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd" Dec 06 09:38:21 crc kubenswrapper[4945]: E1206 09:38:21.245311 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd\": container with ID starting with 8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd not found: ID does not exist" containerID="8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd" Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.245346 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd"} err="failed to get container status \"8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd\": rpc error: code = NotFound desc = could not find container \"8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd\": container with ID starting with 8f8d6ad9a8f2c2305d7695b32d1cbd33ac44ec22656bd4f9ca7e9fc9151942cd not found: ID does not exist" Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.245366 4945 scope.go:117] "RemoveContainer" containerID="2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360" Dec 06 09:38:21 crc kubenswrapper[4945]: E1206 09:38:21.245700 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360\": container with ID starting with 2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360 not found: ID does not exist" containerID="2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360" Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.245725 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360"} err="failed to get container status \"2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360\": rpc error: code = NotFound desc = could not find container \"2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360\": container with ID starting with 2b9888759180ac3abfb8000e0620582aa84f9bc745f7b59f9c3774846de87360 not found: ID does not exist" Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.245742 4945 scope.go:117] "RemoveContainer" containerID="2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4" Dec 06 09:38:21 crc kubenswrapper[4945]: E1206 09:38:21.245960 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4\": container with ID starting with 2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4 not found: ID does not exist" containerID="2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4" Dec 06 09:38:21 crc kubenswrapper[4945]: I1206 09:38:21.245988 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4"} err="failed to get container status \"2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4\": rpc error: code = NotFound desc = could not find container \"2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4\": container with ID starting with 2b043c885fa08aaa790b685fb0f6e5247318e3da43923560b64a15a1320472d4 not found: ID does not exist" Dec 06 09:38:22 crc kubenswrapper[4945]: I1206 09:38:22.965306 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" path="/var/lib/kubelet/pods/db4c1fda-a823-4eb3-9ce9-ee374ec2aefa/volumes" Dec 06 09:39:32 crc kubenswrapper[4945]: I1206 09:39:32.593058 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_5f1d750f-7e4b-487c-a552-5bf7c183c7ad/init-config-reloader/0.log" Dec 06 09:39:32 crc kubenswrapper[4945]: I1206 09:39:32.809846 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_5f1d750f-7e4b-487c-a552-5bf7c183c7ad/init-config-reloader/0.log" Dec 06 09:39:32 crc kubenswrapper[4945]: I1206 09:39:32.829124 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_5f1d750f-7e4b-487c-a552-5bf7c183c7ad/alertmanager/0.log" Dec 06 09:39:33 crc kubenswrapper[4945]: I1206 09:39:33.431118 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_5f1d750f-7e4b-487c-a552-5bf7c183c7ad/config-reloader/0.log" Dec 06 09:39:33 crc kubenswrapper[4945]: I1206 09:39:33.470485 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07/aodh-api/0.log" Dec 06 09:39:33 crc kubenswrapper[4945]: I1206 09:39:33.491897 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07/aodh-evaluator/0.log" Dec 06 09:39:33 crc kubenswrapper[4945]: I1206 09:39:33.630567 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07/aodh-listener/0.log" Dec 06 09:39:33 crc kubenswrapper[4945]: I1206 09:39:33.694606 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_1f4ae1ab-e1e4-4fa4-9ce3-084c09531d07/aodh-notifier/0.log" Dec 06 09:39:33 crc kubenswrapper[4945]: I1206 09:39:33.730969 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-87db876d6-6d2b6_73793c5e-8c22-47c2-b798-76d6f8226d73/barbican-api/0.log" Dec 06 09:39:33 crc kubenswrapper[4945]: I1206 09:39:33.845242 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-87db876d6-6d2b6_73793c5e-8c22-47c2-b798-76d6f8226d73/barbican-api-log/0.log" Dec 06 09:39:34 crc kubenswrapper[4945]: I1206 09:39:34.041658 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-65f7bb6fb8-2265g_430db434-86b0-48d5-a730-86fffbb325e2/barbican-keystone-listener/0.log" Dec 06 09:39:34 crc kubenswrapper[4945]: I1206 09:39:34.240437 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-c6cdc9cfc-2vgms_be9fb1fb-edb3-4bdb-9828-d24b824ee635/barbican-worker/0.log" Dec 06 09:39:34 crc kubenswrapper[4945]: I1206 09:39:34.305929 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-c6cdc9cfc-2vgms_be9fb1fb-edb3-4bdb-9828-d24b824ee635/barbican-worker-log/0.log" Dec 06 09:39:34 crc kubenswrapper[4945]: I1206 09:39:34.362261 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-65f7bb6fb8-2265g_430db434-86b0-48d5-a730-86fffbb325e2/barbican-keystone-listener-log/0.log" Dec 06 09:39:34 crc kubenswrapper[4945]: I1206 09:39:34.527902 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-pmpq5_4eb91f90-eb31-42c6-b632-f4e17e175b17/bootstrap-openstack-openstack-cell1/0.log" Dec 06 09:39:34 crc kubenswrapper[4945]: I1206 09:39:34.640846 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4c910732-ba37-4887-9ab3-32c491361007/ceilometer-central-agent/0.log" Dec 06 09:39:34 crc kubenswrapper[4945]: I1206 09:39:34.740732 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4c910732-ba37-4887-9ab3-32c491361007/ceilometer-notification-agent/0.log" Dec 06 09:39:34 crc kubenswrapper[4945]: I1206 09:39:34.783369 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4c910732-ba37-4887-9ab3-32c491361007/proxy-httpd/0.log" Dec 06 09:39:34 crc kubenswrapper[4945]: I1206 09:39:34.831472 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4c910732-ba37-4887-9ab3-32c491361007/sg-core/0.log" Dec 06 09:39:34 crc kubenswrapper[4945]: I1206 09:39:34.975903 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-w27w6_ac1cf84d-8334-415c-a6cc-3c65a26a7bb2/ceph-client-openstack-openstack-cell1/0.log" Dec 06 09:39:35 crc kubenswrapper[4945]: I1206 09:39:35.140832 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_117f8acf-2899-4353-ad75-8855c73fa61e/cinder-api/0.log" Dec 06 09:39:35 crc kubenswrapper[4945]: I1206 09:39:35.154845 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_117f8acf-2899-4353-ad75-8855c73fa61e/cinder-api-log/0.log" Dec 06 09:39:35 crc kubenswrapper[4945]: I1206 09:39:35.448001 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_737bd8cb-ebc6-441f-9f6c-0f13e356232c/probe/0.log" Dec 06 09:39:35 crc kubenswrapper[4945]: I1206 09:39:35.480457 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_737bd8cb-ebc6-441f-9f6c-0f13e356232c/cinder-backup/0.log" Dec 06 09:39:35 crc kubenswrapper[4945]: I1206 09:39:35.523792 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ed728ae8-6b6b-4edd-b577-91ef23d729e9/cinder-scheduler/0.log" Dec 06 09:39:35 crc kubenswrapper[4945]: I1206 09:39:35.589759 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ed728ae8-6b6b-4edd-b577-91ef23d729e9/probe/0.log" Dec 06 09:39:35 crc kubenswrapper[4945]: I1206 09:39:35.730707 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a15a0964-44c0-4a17-9f3e-42a5ed41f4e9/probe/0.log" Dec 06 09:39:35 crc kubenswrapper[4945]: I1206 09:39:35.732840 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a15a0964-44c0-4a17-9f3e-42a5ed41f4e9/cinder-volume/0.log" Dec 06 09:39:35 crc kubenswrapper[4945]: I1206 09:39:35.838873 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-zddbb_74371052-d51c-4094-be06-38d48cb5053a/configure-network-openstack-openstack-cell1/0.log" Dec 06 09:39:35 crc kubenswrapper[4945]: I1206 09:39:35.995688 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-fxlf6_92e071d6-dcaf-49d5-8dc0-3d8b3264b585/configure-os-openstack-openstack-cell1/0.log" Dec 06 09:39:36 crc kubenswrapper[4945]: I1206 09:39:36.060575 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6fb75d6859-4scbt_92fe4f16-f24e-437c-9f18-1fc3d4f3be78/init/0.log" Dec 06 09:39:36 crc kubenswrapper[4945]: I1206 09:39:36.228577 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6fb75d6859-4scbt_92fe4f16-f24e-437c-9f18-1fc3d4f3be78/init/0.log" Dec 06 09:39:36 crc kubenswrapper[4945]: I1206 09:39:36.276316 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-hhcvg_7775a907-216c-477c-9087-4cd2393ee743/download-cache-openstack-openstack-cell1/0.log" Dec 06 09:39:36 crc kubenswrapper[4945]: I1206 09:39:36.310497 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6fb75d6859-4scbt_92fe4f16-f24e-437c-9f18-1fc3d4f3be78/dnsmasq-dns/0.log" Dec 06 09:39:36 crc kubenswrapper[4945]: I1206 09:39:36.483728 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1f2fce06-f02b-4812-a91d-be28a5b74705/glance-httpd/0.log" Dec 06 09:39:36 crc kubenswrapper[4945]: I1206 09:39:36.533074 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1f2fce06-f02b-4812-a91d-be28a5b74705/glance-log/0.log" Dec 06 09:39:36 crc kubenswrapper[4945]: I1206 09:39:36.607328 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_84ceb1ce-0b26-4b51-bc94-de365bc9849b/glance-httpd/0.log" Dec 06 09:39:36 crc kubenswrapper[4945]: I1206 09:39:36.662013 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_84ceb1ce-0b26-4b51-bc94-de365bc9849b/glance-log/0.log" Dec 06 09:39:36 crc kubenswrapper[4945]: I1206 09:39:36.861863 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-694fddd77-dwqw7_7cd86064-9374-494e-90b6-dac599098554/heat-api/0.log" Dec 06 09:39:36 crc kubenswrapper[4945]: I1206 09:39:36.960807 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-779bcf995f-tvh72_0b92b2c6-fac7-4da7-9291-ffcac487b85f/heat-cfnapi/0.log" Dec 06 09:39:37 crc kubenswrapper[4945]: I1206 09:39:37.040060 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5445fc8644-scpcx_97d626d5-7adf-40dd-9b6a-cf361f3c34cf/heat-engine/0.log" Dec 06 09:39:37 crc kubenswrapper[4945]: I1206 09:39:37.254440 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-gnccs_6fcc9e05-4e5d-434c-bd5b-382d452fc94c/install-certs-openstack-openstack-cell1/0.log" Dec 06 09:39:37 crc kubenswrapper[4945]: I1206 09:39:37.261062 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-64775b6477-kp8ql_e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9/horizon-log/0.log" Dec 06 09:39:37 crc kubenswrapper[4945]: I1206 09:39:37.290993 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-64775b6477-kp8ql_e9022365-1f4a-4b9b-a2d5-32fa96d9c0b9/horizon/0.log" Dec 06 09:39:37 crc kubenswrapper[4945]: I1206 09:39:37.445249 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-vv8q2_c28069dc-d3fb-40e7-8235-72c59c6580d8/install-os-openstack-openstack-cell1/0.log" Dec 06 09:39:37 crc kubenswrapper[4945]: I1206 09:39:37.773551 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29416861-bj6jl_f5127279-51b5-4c93-9ca9-f1d409c09aba/keystone-cron/0.log" Dec 06 09:39:37 crc kubenswrapper[4945]: I1206 09:39:37.856648 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c87cbcc4c-8jnjw_d5e3e5df-4483-49ee-9d22-cf9d3a8757c1/keystone-api/0.log" Dec 06 09:39:37 crc kubenswrapper[4945]: I1206 09:39:37.999814 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_2e198f48-5c2b-4834-9c95-934480712bf2/kube-state-metrics/0.log" Dec 06 09:39:38 crc kubenswrapper[4945]: I1206 09:39:38.030749 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-vh6xt_c1bde47d-052e-4406-86c7-9cf9c8cdf294/libvirt-openstack-openstack-cell1/0.log" Dec 06 09:39:38 crc kubenswrapper[4945]: I1206 09:39:38.223582 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_c07d97d9-2707-4564-86b5-f699f80d5831/manila-api-log/0.log" Dec 06 09:39:38 crc kubenswrapper[4945]: I1206 09:39:38.294907 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_c07d97d9-2707-4564-86b5-f699f80d5831/manila-api/0.log" Dec 06 09:39:38 crc kubenswrapper[4945]: I1206 09:39:38.380163 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f/manila-scheduler/0.log" Dec 06 09:39:38 crc kubenswrapper[4945]: I1206 09:39:38.447681 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_cc24ec7f-ab28-45ed-9ba7-b0e7883aaf6f/probe/0.log" Dec 06 09:39:38 crc kubenswrapper[4945]: I1206 09:39:38.565944 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b19fcd2e-e77a-4543-ac62-5ecef9e48b7c/manila-share/0.log" Dec 06 09:39:38 crc kubenswrapper[4945]: I1206 09:39:38.589828 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b19fcd2e-e77a-4543-ac62-5ecef9e48b7c/probe/0.log" Dec 06 09:39:39 crc kubenswrapper[4945]: I1206 09:39:39.021202 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d89777897-g2pzk_17bb80e7-5bad-496c-a404-04f4034de941/neutron-httpd/0.log" Dec 06 09:39:39 crc kubenswrapper[4945]: I1206 09:39:39.093351 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-whzbj_94f979eb-bec9-48f7-b87f-f9618e999665/neutron-dhcp-openstack-openstack-cell1/0.log" Dec 06 09:39:39 crc kubenswrapper[4945]: I1206 09:39:39.333537 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-5xr9h_8771b333-f1cc-4655-b6af-8e7eddcb2b7a/neutron-metadata-openstack-openstack-cell1/0.log" Dec 06 09:39:39 crc kubenswrapper[4945]: I1206 09:39:39.459478 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d89777897-g2pzk_17bb80e7-5bad-496c-a404-04f4034de941/neutron-api/0.log" Dec 06 09:39:39 crc kubenswrapper[4945]: I1206 09:39:39.557395 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-x45jt_831dbf28-be96-4229-bd44-f88bb5b34e9c/neutron-sriov-openstack-openstack-cell1/0.log" Dec 06 09:39:39 crc kubenswrapper[4945]: I1206 09:39:39.908838 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0ec077cb-3c84-4617-bd63-6333112b514e/nova-api-api/0.log" Dec 06 09:39:39 crc kubenswrapper[4945]: I1206 09:39:39.997489 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0ec077cb-3c84-4617-bd63-6333112b514e/nova-api-log/0.log" Dec 06 09:39:40 crc kubenswrapper[4945]: I1206 09:39:40.028114 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_0e477bf3-d8bf-47ae-9b81-080d365397ac/nova-cell0-conductor-conductor/0.log" Dec 06 09:39:40 crc kubenswrapper[4945]: I1206 09:39:40.182889 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_9ef2d829-66ae-404a-9802-f7c696bf8d7d/nova-cell1-conductor-conductor/0.log" Dec 06 09:39:40 crc kubenswrapper[4945]: I1206 09:39:40.387756 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f53d6629-9d10-4331-8f9f-ef26361c7e3b/nova-cell1-novncproxy-novncproxy/0.log" Dec 06 09:39:40 crc kubenswrapper[4945]: I1206 09:39:40.511430 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellbtlwm_c5c4bb2e-c6c7-49f4-8c53-32e4dba43b4c/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Dec 06 09:39:40 crc kubenswrapper[4945]: I1206 09:39:40.717126 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-4pkmq_d03da9d8-fe0b-4894-806d-2d0f15de1ea1/nova-cell1-openstack-openstack-cell1/0.log" Dec 06 09:39:40 crc kubenswrapper[4945]: I1206 09:39:40.821510 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_18b0940d-3e8f-4d9d-ab7c-530e5da63e4d/nova-metadata-log/0.log" Dec 06 09:39:40 crc kubenswrapper[4945]: I1206 09:39:40.953878 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_18b0940d-3e8f-4d9d-ab7c-530e5da63e4d/nova-metadata-metadata/0.log" Dec 06 09:39:41 crc kubenswrapper[4945]: I1206 09:39:41.132681 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_5e224dc5-9c91-4441-8634-3e43c8cf89e0/nova-scheduler-scheduler/0.log" Dec 06 09:39:41 crc kubenswrapper[4945]: I1206 09:39:41.171078 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0ea9657b-8247-4130-bbd7-f8c4b2f12b50/mysql-bootstrap/0.log" Dec 06 09:39:41 crc kubenswrapper[4945]: I1206 09:39:41.731642 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0ea9657b-8247-4130-bbd7-f8c4b2f12b50/mysql-bootstrap/0.log" Dec 06 09:39:41 crc kubenswrapper[4945]: I1206 09:39:41.807024 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_adf8ff2c-2982-4a81-899e-12d36a3c0fef/mysql-bootstrap/0.log" Dec 06 09:39:41 crc kubenswrapper[4945]: I1206 09:39:41.842302 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0ea9657b-8247-4130-bbd7-f8c4b2f12b50/galera/0.log" Dec 06 09:39:42 crc kubenswrapper[4945]: I1206 09:39:42.062711 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_adf8ff2c-2982-4a81-899e-12d36a3c0fef/mysql-bootstrap/0.log" Dec 06 09:39:42 crc kubenswrapper[4945]: I1206 09:39:42.142398 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_d7b129e8-fa9c-47f4-bb42-ed7867d79609/openstackclient/0.log" Dec 06 09:39:42 crc kubenswrapper[4945]: I1206 09:39:42.161787 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_adf8ff2c-2982-4a81-899e-12d36a3c0fef/galera/0.log" Dec 06 09:39:42 crc kubenswrapper[4945]: I1206 09:39:42.368430 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ba9377fd-f829-45bb-8478-a233f5cb6a81/openstack-network-exporter/0.log" Dec 06 09:39:42 crc kubenswrapper[4945]: I1206 09:39:42.373562 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ba9377fd-f829-45bb-8478-a233f5cb6a81/ovn-northd/0.log" Dec 06 09:39:42 crc kubenswrapper[4945]: I1206 09:39:42.608950 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_befb328f-40fa-41b6-845e-79c19af5ce49/openstack-network-exporter/0.log" Dec 06 09:39:42 crc kubenswrapper[4945]: I1206 09:39:42.661838 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-j7cpj_4bbbd623-f02c-4b7b-ab9c-f33e570a7b02/ovn-openstack-openstack-cell1/0.log" Dec 06 09:39:42 crc kubenswrapper[4945]: I1206 09:39:42.739248 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_befb328f-40fa-41b6-845e-79c19af5ce49/ovsdbserver-nb/0.log" Dec 06 09:39:42 crc kubenswrapper[4945]: I1206 09:39:42.863600 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_41bd1596-3908-44e6-840c-3011459fdc81/openstack-network-exporter/0.log" Dec 06 09:39:42 crc kubenswrapper[4945]: I1206 09:39:42.934257 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_41bd1596-3908-44e6-840c-3011459fdc81/ovsdbserver-nb/0.log" Dec 06 09:39:43 crc kubenswrapper[4945]: I1206 09:39:43.065263 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_62420d57-1545-4d7e-834f-60d8c6a6a4ff/openstack-network-exporter/0.log" Dec 06 09:39:43 crc kubenswrapper[4945]: I1206 09:39:43.070149 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_62420d57-1545-4d7e-834f-60d8c6a6a4ff/ovsdbserver-nb/0.log" Dec 06 09:39:43 crc kubenswrapper[4945]: I1206 09:39:43.625414 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b14a3e86-54c1-46bd-b470-3ae5e648b2ff/openstack-network-exporter/0.log" Dec 06 09:39:43 crc kubenswrapper[4945]: I1206 09:39:43.687306 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b14a3e86-54c1-46bd-b470-3ae5e648b2ff/ovsdbserver-sb/0.log" Dec 06 09:39:43 crc kubenswrapper[4945]: I1206 09:39:43.834072 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_744a0d22-5718-4d6a-ad18-2f522e666977/openstack-network-exporter/0.log" Dec 06 09:39:43 crc kubenswrapper[4945]: I1206 09:39:43.908875 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_744a0d22-5718-4d6a-ad18-2f522e666977/ovsdbserver-sb/0.log" Dec 06 09:39:43 crc kubenswrapper[4945]: I1206 09:39:43.976385 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_97504023-3ba7-4480-9c54-ff4985c3f38c/openstack-network-exporter/0.log" Dec 06 09:39:44 crc kubenswrapper[4945]: I1206 09:39:44.063033 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_97504023-3ba7-4480-9c54-ff4985c3f38c/ovsdbserver-sb/0.log" Dec 06 09:39:44 crc kubenswrapper[4945]: I1206 09:39:44.405256 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5c8c7ff5f6-w4pdr_8c5cb978-dca7-4d06-8c84-c06588c6eacd/placement-log/0.log" Dec 06 09:39:44 crc kubenswrapper[4945]: I1206 09:39:44.405528 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5c8c7ff5f6-w4pdr_8c5cb978-dca7-4d06-8c84-c06588c6eacd/placement-api/0.log" Dec 06 09:39:44 crc kubenswrapper[4945]: I1206 09:39:44.406897 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-chr88c_7f91cf82-922a-490d-861d-f1914f61b7da/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Dec 06 09:39:44 crc kubenswrapper[4945]: I1206 09:39:44.617476 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_4c654f23-0692-4d36-918c-3e669d026c44/init-config-reloader/0.log" Dec 06 09:39:44 crc kubenswrapper[4945]: I1206 09:39:44.823478 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_4c654f23-0692-4d36-918c-3e669d026c44/config-reloader/0.log" Dec 06 09:39:44 crc kubenswrapper[4945]: I1206 09:39:44.831996 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_4c654f23-0692-4d36-918c-3e669d026c44/init-config-reloader/0.log" Dec 06 09:39:44 crc kubenswrapper[4945]: I1206 09:39:44.871666 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_4c654f23-0692-4d36-918c-3e669d026c44/thanos-sidecar/0.log" Dec 06 09:39:44 crc kubenswrapper[4945]: I1206 09:39:44.943563 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_4c654f23-0692-4d36-918c-3e669d026c44/prometheus/0.log" Dec 06 09:39:45 crc kubenswrapper[4945]: I1206 09:39:45.064504 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e36f4855-c747-4b0e-8648-3b7619707dcb/setup-container/0.log" Dec 06 09:39:45 crc kubenswrapper[4945]: I1206 09:39:45.315536 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e36f4855-c747-4b0e-8648-3b7619707dcb/setup-container/0.log" Dec 06 09:39:45 crc kubenswrapper[4945]: I1206 09:39:45.354984 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e36f4855-c747-4b0e-8648-3b7619707dcb/rabbitmq/0.log" Dec 06 09:39:45 crc kubenswrapper[4945]: I1206 09:39:45.404553 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc/setup-container/0.log" Dec 06 09:39:45 crc kubenswrapper[4945]: I1206 09:39:45.591712 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc/setup-container/0.log" Dec 06 09:39:45 crc kubenswrapper[4945]: I1206 09:39:45.727639 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-bkpcr_74f5fed5-012c-44bd-8e9c-4576689abe4c/reboot-os-openstack-openstack-cell1/0.log" Dec 06 09:39:45 crc kubenswrapper[4945]: I1206 09:39:45.906253 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-qtb45_fc8f860e-c526-4d6a-b91b-6a6ed10e176b/run-os-openstack-openstack-cell1/0.log" Dec 06 09:39:46 crc kubenswrapper[4945]: I1206 09:39:46.148670 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-skqg9_aa10039e-9b12-42f0-bdf9-b91b3e7acf10/ssh-known-hosts-openstack/0.log" Dec 06 09:39:46 crc kubenswrapper[4945]: I1206 09:39:46.361787 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9a2c4c88-0eb3-40d9-923e-8e6f2f05dedc/rabbitmq/0.log" Dec 06 09:39:46 crc kubenswrapper[4945]: I1206 09:39:46.424547 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-rlgng_373f7931-312f-4269-932e-53fc9a2e1e6d/telemetry-openstack-openstack-cell1/0.log" Dec 06 09:39:46 crc kubenswrapper[4945]: I1206 09:39:46.461522 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1000996d-89a3-43f6-bffc-03fd64593568/tempest-tests-tempest-tests-runner/0.log" Dec 06 09:39:46 crc kubenswrapper[4945]: I1206 09:39:46.647023 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_ba384fba-e2b7-4e9b-b16b-28673c1567f9/test-operator-logs-container/0.log" Dec 06 09:39:46 crc kubenswrapper[4945]: I1206 09:39:46.694604 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-8dhjq_73966546-b73a-4608-ab6c-ab094287f285/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Dec 06 09:39:46 crc kubenswrapper[4945]: I1206 09:39:46.912605 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-gch6j_653aa8ac-b5b0-47a6-9901-f6d09ca71674/validate-network-openstack-openstack-cell1/0.log" Dec 06 09:40:05 crc kubenswrapper[4945]: I1206 09:40:05.623579 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2c37a3a7-2c98-45ce-b22a-2ea765bf240e/memcached/0.log" Dec 06 09:40:12 crc kubenswrapper[4945]: I1206 09:40:12.434064 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx_9329d400-df2a-48e9-bdee-cdaa2304ae0e/util/0.log" Dec 06 09:40:12 crc kubenswrapper[4945]: I1206 09:40:12.627554 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx_9329d400-df2a-48e9-bdee-cdaa2304ae0e/util/0.log" Dec 06 09:40:12 crc kubenswrapper[4945]: I1206 09:40:12.631762 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx_9329d400-df2a-48e9-bdee-cdaa2304ae0e/pull/0.log" Dec 06 09:40:12 crc kubenswrapper[4945]: I1206 09:40:12.678989 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx_9329d400-df2a-48e9-bdee-cdaa2304ae0e/pull/0.log" Dec 06 09:40:12 crc kubenswrapper[4945]: I1206 09:40:12.836030 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx_9329d400-df2a-48e9-bdee-cdaa2304ae0e/extract/0.log" Dec 06 09:40:12 crc kubenswrapper[4945]: I1206 09:40:12.853230 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx_9329d400-df2a-48e9-bdee-cdaa2304ae0e/util/0.log" Dec 06 09:40:12 crc kubenswrapper[4945]: I1206 09:40:12.906523 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaflkjgx_9329d400-df2a-48e9-bdee-cdaa2304ae0e/pull/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.045210 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-wqq5g_7305df73-ea7d-43f5-9119-9c12942796bc/kube-rbac-proxy/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.126780 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-qd55f_afd67b97-8d07-4522-b37d-019abf73ee85/kube-rbac-proxy/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.262909 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-wqq5g_7305df73-ea7d-43f5-9119-9c12942796bc/manager/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.349838 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-qd55f_afd67b97-8d07-4522-b37d-019abf73ee85/manager/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.387702 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-kjxk4_31beec86-b14f-43c6-9307-a776454ed9b7/kube-rbac-proxy/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.485412 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-kjxk4_31beec86-b14f-43c6-9307-a776454ed9b7/manager/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.614035 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-cgbvh_c925c68e-4099-403e-b8ba-84d2d9bc689e/kube-rbac-proxy/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.738831 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-n8gdh_8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd/kube-rbac-proxy/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.756495 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-cgbvh_c925c68e-4099-403e-b8ba-84d2d9bc689e/manager/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.922840 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-n8gdh_8f2cb38d-eec1-4ba8-a0a5-fc3bcece7afd/manager/0.log" Dec 06 09:40:13 crc kubenswrapper[4945]: I1206 09:40:13.986426 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-cchx8_f32f27d3-90de-4c9e-8664-68e149e42ffc/kube-rbac-proxy/0.log" Dec 06 09:40:14 crc kubenswrapper[4945]: I1206 09:40:14.035621 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-cchx8_f32f27d3-90de-4c9e-8664-68e149e42ffc/manager/0.log" Dec 06 09:40:14 crc kubenswrapper[4945]: I1206 09:40:14.221820 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-qhvjz_db9c0007-43e1-4cb2-ae8a-44a4731254b7/kube-rbac-proxy/0.log" Dec 06 09:40:14 crc kubenswrapper[4945]: I1206 09:40:14.361247 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-r7zm6_b89bffc9-f64a-4d0b-b5e5-104db0d681dd/kube-rbac-proxy/0.log" Dec 06 09:40:14 crc kubenswrapper[4945]: I1206 09:40:14.501458 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-r7zm6_b89bffc9-f64a-4d0b-b5e5-104db0d681dd/manager/0.log" Dec 06 09:40:14 crc kubenswrapper[4945]: I1206 09:40:14.583614 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-qhvjz_db9c0007-43e1-4cb2-ae8a-44a4731254b7/manager/0.log" Dec 06 09:40:14 crc kubenswrapper[4945]: I1206 09:40:14.646075 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-x98dd_7c3f47b7-f1f2-408d-80d0-03a0ea9286ea/kube-rbac-proxy/0.log" Dec 06 09:40:14 crc kubenswrapper[4945]: I1206 09:40:14.805589 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-x98dd_7c3f47b7-f1f2-408d-80d0-03a0ea9286ea/manager/0.log" Dec 06 09:40:14 crc kubenswrapper[4945]: I1206 09:40:14.837718 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-p2gzt_40318b4d-ed4f-478b-8333-8e1cb0729b34/kube-rbac-proxy/0.log" Dec 06 09:40:14 crc kubenswrapper[4945]: I1206 09:40:14.943738 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-p2gzt_40318b4d-ed4f-478b-8333-8e1cb0729b34/manager/0.log" Dec 06 09:40:15 crc kubenswrapper[4945]: I1206 09:40:15.062320 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-f6rz7_3c4830bb-bc9c-447e-bb30-9c481fae8234/kube-rbac-proxy/0.log" Dec 06 09:40:15 crc kubenswrapper[4945]: I1206 09:40:15.123612 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-f6rz7_3c4830bb-bc9c-447e-bb30-9c481fae8234/manager/0.log" Dec 06 09:40:15 crc kubenswrapper[4945]: I1206 09:40:15.279242 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-5njm2_0e92db58-a1fc-4035-b635-f274d7411813/kube-rbac-proxy/0.log" Dec 06 09:40:15 crc kubenswrapper[4945]: I1206 09:40:15.392440 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-5njm2_0e92db58-a1fc-4035-b635-f274d7411813/manager/0.log" Dec 06 09:40:15 crc kubenswrapper[4945]: I1206 09:40:15.535939 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-bpsxr_6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc/kube-rbac-proxy/0.log" Dec 06 09:40:15 crc kubenswrapper[4945]: I1206 09:40:15.689219 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-pd42f_9bfd7a39-d83d-42de-bb30-0d1dc17bedca/kube-rbac-proxy/0.log" Dec 06 09:40:15 crc kubenswrapper[4945]: I1206 09:40:15.699100 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-bpsxr_6d50597e-adf1-4ce6-8f1c-f2acfcbbd0cc/manager/0.log" Dec 06 09:40:15 crc kubenswrapper[4945]: I1206 09:40:15.807424 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-pd42f_9bfd7a39-d83d-42de-bb30-0d1dc17bedca/manager/0.log" Dec 06 09:40:15 crc kubenswrapper[4945]: I1206 09:40:15.928566 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55c85496f56d49f_19942e24-acd1-469c-94b2-37063e88c190/kube-rbac-proxy/0.log" Dec 06 09:40:15 crc kubenswrapper[4945]: I1206 09:40:15.938686 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55c85496f56d49f_19942e24-acd1-469c-94b2-37063e88c190/manager/0.log" Dec 06 09:40:16 crc kubenswrapper[4945]: I1206 09:40:16.473952 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55b6fb9447-lfnj5_5360a460-56fc-4dd6-9014-614ed5a8e05e/operator/0.log" Dec 06 09:40:16 crc kubenswrapper[4945]: I1206 09:40:16.582741 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-5h9ss_5da80984-5ba5-48be-b853-847ae00cba6c/registry-server/0.log" Dec 06 09:40:17 crc kubenswrapper[4945]: I1206 09:40:17.286250 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lwkxm_6c628f21-0f62-4761-9dd1-f6377a5633a5/kube-rbac-proxy/0.log" Dec 06 09:40:17 crc kubenswrapper[4945]: I1206 09:40:17.532116 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-7cn4x_039d67fe-f691-49a7-8e38-4a8dc2a2227e/kube-rbac-proxy/0.log" Dec 06 09:40:17 crc kubenswrapper[4945]: I1206 09:40:17.569815 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lwkxm_6c628f21-0f62-4761-9dd1-f6377a5633a5/manager/0.log" Dec 06 09:40:17 crc kubenswrapper[4945]: I1206 09:40:17.621952 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-7cn4x_039d67fe-f691-49a7-8e38-4a8dc2a2227e/manager/0.log" Dec 06 09:40:17 crc kubenswrapper[4945]: I1206 09:40:17.861790 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-bd5w8_4fe7c3eb-89a4-4935-8e19-935f98f8562b/operator/0.log" Dec 06 09:40:17 crc kubenswrapper[4945]: I1206 09:40:17.869799 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6tvc7_716df590-6583-4b4b-bfdc-4d6bb2ac05a4/kube-rbac-proxy/0.log" Dec 06 09:40:17 crc kubenswrapper[4945]: I1206 09:40:17.902272 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6tvc7_716df590-6583-4b4b-bfdc-4d6bb2ac05a4/manager/0.log" Dec 06 09:40:18 crc kubenswrapper[4945]: I1206 09:40:18.107420 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-c4q9q_4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4/kube-rbac-proxy/0.log" Dec 06 09:40:18 crc kubenswrapper[4945]: I1206 09:40:18.371967 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-c4q9q_4c5c1b2c-d0aa-4dd7-afc9-9cd805c7a2b4/manager/0.log" Dec 06 09:40:18 crc kubenswrapper[4945]: I1206 09:40:18.614165 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-54bdf956c4-f6b9t_ad774804-402d-45d6-957f-d7ab18bd2f7f/manager/0.log" Dec 06 09:40:18 crc kubenswrapper[4945]: I1206 09:40:18.900791 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-v5krd_f19785e5-2908-48b0-aa6e-ea4f37bbd506/kube-rbac-proxy/0.log" Dec 06 09:40:18 crc kubenswrapper[4945]: I1206 09:40:18.998307 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-v5krd_f19785e5-2908-48b0-aa6e-ea4f37bbd506/manager/0.log" Dec 06 09:40:19 crc kubenswrapper[4945]: I1206 09:40:19.012637 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-2vzqx_c0722d35-7779-4073-851f-0171050b6141/kube-rbac-proxy/0.log" Dec 06 09:40:19 crc kubenswrapper[4945]: I1206 09:40:19.057357 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-2vzqx_c0722d35-7779-4073-851f-0171050b6141/manager/0.log" Dec 06 09:40:38 crc kubenswrapper[4945]: I1206 09:40:38.795989 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:40:38 crc kubenswrapper[4945]: I1206 09:40:38.796660 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:40:40 crc kubenswrapper[4945]: I1206 09:40:40.021644 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jv77f_23ec3b33-0492-49bd-b1df-aa5aaac01315/control-plane-machine-set-operator/0.log" Dec 06 09:40:40 crc kubenswrapper[4945]: I1206 09:40:40.246414 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4qjdx_4e308683-ec1e-400b-8503-e671d8ac4589/machine-api-operator/0.log" Dec 06 09:40:40 crc kubenswrapper[4945]: I1206 09:40:40.257135 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4qjdx_4e308683-ec1e-400b-8503-e671d8ac4589/kube-rbac-proxy/0.log" Dec 06 09:40:52 crc kubenswrapper[4945]: I1206 09:40:52.940164 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-cpjql_8ca8d146-bf33-4f3e-94e2-a794187ee78e/cert-manager-controller/0.log" Dec 06 09:40:53 crc kubenswrapper[4945]: I1206 09:40:53.140523 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-qh9wt_02a40739-7cbd-4492-aff7-ab1dd5bd63c8/cert-manager-cainjector/0.log" Dec 06 09:40:53 crc kubenswrapper[4945]: I1206 09:40:53.234043 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-ms6cl_96f7adf5-11a3-480d-ae16-cb4bf53ab693/cert-manager-webhook/0.log" Dec 06 09:41:05 crc kubenswrapper[4945]: I1206 09:41:05.677738 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-lrsrq_49085c7f-2b2f-4692-a76a-b1a159f8e121/nmstate-console-plugin/0.log" Dec 06 09:41:05 crc kubenswrapper[4945]: I1206 09:41:05.866912 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-lb98q_ae31a742-6313-4d6c-bfe2-989f1e46d401/nmstate-handler/0.log" Dec 06 09:41:05 crc kubenswrapper[4945]: I1206 09:41:05.957726 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-4999s_fcf83bdc-5289-4109-af7f-10c3198eac36/kube-rbac-proxy/0.log" Dec 06 09:41:05 crc kubenswrapper[4945]: I1206 09:41:05.982821 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-4999s_fcf83bdc-5289-4109-af7f-10c3198eac36/nmstate-metrics/0.log" Dec 06 09:41:06 crc kubenswrapper[4945]: I1206 09:41:06.146318 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-s72f2_bfdca761-893a-4fb8-b500-af0d0bdf44e8/nmstate-operator/0.log" Dec 06 09:41:06 crc kubenswrapper[4945]: I1206 09:41:06.192577 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-4kdsq_9fc802d9-93bc-46af-acba-93258bfb4890/nmstate-webhook/0.log" Dec 06 09:41:08 crc kubenswrapper[4945]: I1206 09:41:08.796625 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:41:08 crc kubenswrapper[4945]: I1206 09:41:08.797509 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:41:21 crc kubenswrapper[4945]: I1206 09:41:21.226546 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-m6c5w_98251a38-5e6f-4fff-b389-1ffec3864d22/kube-rbac-proxy/0.log" Dec 06 09:41:21 crc kubenswrapper[4945]: I1206 09:41:21.619698 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-m6c5w_98251a38-5e6f-4fff-b389-1ffec3864d22/controller/0.log" Dec 06 09:41:21 crc kubenswrapper[4945]: I1206 09:41:21.835448 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-frr-files/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.046496 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-metrics/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.098356 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-reloader/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.099863 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-reloader/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.100393 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-frr-files/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.309722 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-reloader/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.312662 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-frr-files/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.330153 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-metrics/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.339336 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-metrics/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.508007 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-frr-files/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.537380 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-metrics/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.554212 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/cp-reloader/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.585861 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/controller/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.714485 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/frr-metrics/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.731447 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/kube-rbac-proxy/0.log" Dec 06 09:41:22 crc kubenswrapper[4945]: I1206 09:41:22.819513 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/kube-rbac-proxy-frr/0.log" Dec 06 09:41:23 crc kubenswrapper[4945]: I1206 09:41:23.010767 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-6ljkh_a2254f20-d815-4f62-98e5-3fb4bd20fd49/frr-k8s-webhook-server/0.log" Dec 06 09:41:23 crc kubenswrapper[4945]: I1206 09:41:23.108646 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/reloader/0.log" Dec 06 09:41:23 crc kubenswrapper[4945]: I1206 09:41:23.304224 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6dd5c5658d-plhqt_ae4cf098-3915-418a-ae09-e1d08bfc523b/manager/0.log" Dec 06 09:41:23 crc kubenswrapper[4945]: I1206 09:41:23.553912 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-685db896b9-sn7lp_30d3f7c2-7fc1-463c-9cf7-f7f6b2f74ab0/webhook-server/0.log" Dec 06 09:41:23 crc kubenswrapper[4945]: I1206 09:41:23.644523 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-pfl8t_7e84c3f8-86bc-4ce1-9e48-936734a78b6d/kube-rbac-proxy/0.log" Dec 06 09:41:24 crc kubenswrapper[4945]: I1206 09:41:24.481318 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-pfl8t_7e84c3f8-86bc-4ce1-9e48-936734a78b6d/speaker/0.log" Dec 06 09:41:25 crc kubenswrapper[4945]: I1206 09:41:25.958516 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-269rv_58b9a31e-1915-40f7-8e45-cca0242826aa/frr/0.log" Dec 06 09:41:37 crc kubenswrapper[4945]: I1206 09:41:37.286960 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44_0549fa78-72df-422e-b112-cdff8788ebcb/util/0.log" Dec 06 09:41:37 crc kubenswrapper[4945]: I1206 09:41:37.438981 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44_0549fa78-72df-422e-b112-cdff8788ebcb/util/0.log" Dec 06 09:41:37 crc kubenswrapper[4945]: I1206 09:41:37.460257 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44_0549fa78-72df-422e-b112-cdff8788ebcb/pull/0.log" Dec 06 09:41:37 crc kubenswrapper[4945]: I1206 09:41:37.506288 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44_0549fa78-72df-422e-b112-cdff8788ebcb/pull/0.log" Dec 06 09:41:37 crc kubenswrapper[4945]: I1206 09:41:37.639600 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44_0549fa78-72df-422e-b112-cdff8788ebcb/util/0.log" Dec 06 09:41:37 crc kubenswrapper[4945]: I1206 09:41:37.652030 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44_0549fa78-72df-422e-b112-cdff8788ebcb/pull/0.log" Dec 06 09:41:37 crc kubenswrapper[4945]: I1206 09:41:37.673934 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4mb44_0549fa78-72df-422e-b112-cdff8788ebcb/extract/0.log" Dec 06 09:41:37 crc kubenswrapper[4945]: I1206 09:41:37.797490 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd_466623ca-b9e9-4bf2-82e9-9049583349b7/util/0.log" Dec 06 09:41:37 crc kubenswrapper[4945]: I1206 09:41:37.998893 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd_466623ca-b9e9-4bf2-82e9-9049583349b7/pull/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.024118 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd_466623ca-b9e9-4bf2-82e9-9049583349b7/util/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.030432 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd_466623ca-b9e9-4bf2-82e9-9049583349b7/pull/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.214832 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd_466623ca-b9e9-4bf2-82e9-9049583349b7/util/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.215362 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd_466623ca-b9e9-4bf2-82e9-9049583349b7/extract/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.219024 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgcjdd_466623ca-b9e9-4bf2-82e9-9049583349b7/pull/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.377301 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl_e6cd545a-16b7-49ca-bcb9-3a733c7f4499/util/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.572614 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl_e6cd545a-16b7-49ca-bcb9-3a733c7f4499/pull/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.596918 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl_e6cd545a-16b7-49ca-bcb9-3a733c7f4499/util/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.597556 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl_e6cd545a-16b7-49ca-bcb9-3a733c7f4499/pull/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.784870 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl_e6cd545a-16b7-49ca-bcb9-3a733c7f4499/pull/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.795494 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.795724 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.795839 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.796424 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"64748f113951c547ded8dabe4d2c5762e6623dd8ea101b8b8c6080bfda5b945a"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.796567 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://64748f113951c547ded8dabe4d2c5762e6623dd8ea101b8b8c6080bfda5b945a" gracePeriod=600 Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.831715 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl_e6cd545a-16b7-49ca-bcb9-3a733c7f4499/util/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.851049 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921048vsl_e6cd545a-16b7-49ca-bcb9-3a733c7f4499/extract/0.log" Dec 06 09:41:38 crc kubenswrapper[4945]: I1206 09:41:38.982532 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx_8168c399-bb2d-4c71-ae91-d96758be7a42/util/0.log" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.025780 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="64748f113951c547ded8dabe4d2c5762e6623dd8ea101b8b8c6080bfda5b945a" exitCode=0 Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.025829 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"64748f113951c547ded8dabe4d2c5762e6623dd8ea101b8b8c6080bfda5b945a"} Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.025894 4945 scope.go:117] "RemoveContainer" containerID="0876dd8b933b2d77ef51fe6b9c651428546c4480cf77c015000964812e1f823f" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.191090 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx_8168c399-bb2d-4c71-ae91-d96758be7a42/pull/0.log" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.191174 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx_8168c399-bb2d-4c71-ae91-d96758be7a42/pull/0.log" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.228342 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx_8168c399-bb2d-4c71-ae91-d96758be7a42/util/0.log" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.367479 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx_8168c399-bb2d-4c71-ae91-d96758be7a42/util/0.log" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.383189 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx_8168c399-bb2d-4c71-ae91-d96758be7a42/pull/0.log" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.408420 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83tvpgx_8168c399-bb2d-4c71-ae91-d96758be7a42/extract/0.log" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.545782 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rcv8f_417e92a7-1574-4c0a-ae51-c11449157595/extract-utilities/0.log" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.758003 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rcv8f_417e92a7-1574-4c0a-ae51-c11449157595/extract-content/0.log" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.758937 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rcv8f_417e92a7-1574-4c0a-ae51-c11449157595/extract-utilities/0.log" Dec 06 09:41:39 crc kubenswrapper[4945]: I1206 09:41:39.803487 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rcv8f_417e92a7-1574-4c0a-ae51-c11449157595/extract-content/0.log" Dec 06 09:41:40 crc kubenswrapper[4945]: I1206 09:41:40.004400 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rcv8f_417e92a7-1574-4c0a-ae51-c11449157595/extract-utilities/0.log" Dec 06 09:41:40 crc kubenswrapper[4945]: I1206 09:41:40.038704 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerStarted","Data":"50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931"} Dec 06 09:41:40 crc kubenswrapper[4945]: I1206 09:41:40.086662 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rcv8f_417e92a7-1574-4c0a-ae51-c11449157595/extract-content/0.log" Dec 06 09:41:40 crc kubenswrapper[4945]: I1206 09:41:40.286209 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h8kzt_09e6adf9-3142-4984-8ee3-02ab261667e6/extract-utilities/0.log" Dec 06 09:41:40 crc kubenswrapper[4945]: I1206 09:41:40.508536 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h8kzt_09e6adf9-3142-4984-8ee3-02ab261667e6/extract-utilities/0.log" Dec 06 09:41:40 crc kubenswrapper[4945]: I1206 09:41:40.516046 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h8kzt_09e6adf9-3142-4984-8ee3-02ab261667e6/extract-content/0.log" Dec 06 09:41:40 crc kubenswrapper[4945]: I1206 09:41:40.611612 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h8kzt_09e6adf9-3142-4984-8ee3-02ab261667e6/extract-content/0.log" Dec 06 09:41:40 crc kubenswrapper[4945]: I1206 09:41:40.787362 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h8kzt_09e6adf9-3142-4984-8ee3-02ab261667e6/extract-content/0.log" Dec 06 09:41:40 crc kubenswrapper[4945]: I1206 09:41:40.924304 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h8kzt_09e6adf9-3142-4984-8ee3-02ab261667e6/extract-utilities/0.log" Dec 06 09:41:41 crc kubenswrapper[4945]: I1206 09:41:41.196541 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rgp5_5d136274-fa83-4411-845d-5d5cc3cff1b7/marketplace-operator/0.log" Dec 06 09:41:41 crc kubenswrapper[4945]: I1206 09:41:41.301636 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rcv8f_417e92a7-1574-4c0a-ae51-c11449157595/registry-server/0.log" Dec 06 09:41:41 crc kubenswrapper[4945]: I1206 09:41:41.334989 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gx58r_95ca28b7-42c4-4af4-8b14-d818ef8565e3/extract-utilities/0.log" Dec 06 09:41:41 crc kubenswrapper[4945]: I1206 09:41:41.930482 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gx58r_95ca28b7-42c4-4af4-8b14-d818ef8565e3/extract-utilities/0.log" Dec 06 09:41:41 crc kubenswrapper[4945]: I1206 09:41:41.950792 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gx58r_95ca28b7-42c4-4af4-8b14-d818ef8565e3/extract-content/0.log" Dec 06 09:41:41 crc kubenswrapper[4945]: I1206 09:41:41.996151 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gx58r_95ca28b7-42c4-4af4-8b14-d818ef8565e3/extract-content/0.log" Dec 06 09:41:42 crc kubenswrapper[4945]: I1206 09:41:42.141997 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gx58r_95ca28b7-42c4-4af4-8b14-d818ef8565e3/extract-utilities/0.log" Dec 06 09:41:42 crc kubenswrapper[4945]: I1206 09:41:42.192545 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h8kzt_09e6adf9-3142-4984-8ee3-02ab261667e6/registry-server/0.log" Dec 06 09:41:42 crc kubenswrapper[4945]: I1206 09:41:42.201964 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gx58r_95ca28b7-42c4-4af4-8b14-d818ef8565e3/extract-content/0.log" Dec 06 09:41:42 crc kubenswrapper[4945]: I1206 09:41:42.353219 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xgbc5_a19579b9-98ab-40de-b39a-9419031a180c/extract-utilities/0.log" Dec 06 09:41:42 crc kubenswrapper[4945]: I1206 09:41:42.523815 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gx58r_95ca28b7-42c4-4af4-8b14-d818ef8565e3/registry-server/0.log" Dec 06 09:41:42 crc kubenswrapper[4945]: I1206 09:41:42.561782 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xgbc5_a19579b9-98ab-40de-b39a-9419031a180c/extract-content/0.log" Dec 06 09:41:42 crc kubenswrapper[4945]: I1206 09:41:42.581832 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xgbc5_a19579b9-98ab-40de-b39a-9419031a180c/extract-utilities/0.log" Dec 06 09:41:42 crc kubenswrapper[4945]: I1206 09:41:42.599481 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xgbc5_a19579b9-98ab-40de-b39a-9419031a180c/extract-content/0.log" Dec 06 09:41:42 crc kubenswrapper[4945]: I1206 09:41:42.745414 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xgbc5_a19579b9-98ab-40de-b39a-9419031a180c/extract-utilities/0.log" Dec 06 09:41:42 crc kubenswrapper[4945]: I1206 09:41:42.747883 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xgbc5_a19579b9-98ab-40de-b39a-9419031a180c/extract-content/0.log" Dec 06 09:41:43 crc kubenswrapper[4945]: I1206 09:41:43.992232 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xgbc5_a19579b9-98ab-40de-b39a-9419031a180c/registry-server/0.log" Dec 06 09:41:56 crc kubenswrapper[4945]: I1206 09:41:56.576802 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-kmwcv_37a5d5bb-936f-4c5b-a808-e88fc434aa2f/prometheus-operator/0.log" Dec 06 09:41:56 crc kubenswrapper[4945]: I1206 09:41:56.671995 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-546584f4f9-4wm9k_19169e62-1838-4277-a99d-f9b2944e37c2/prometheus-operator-admission-webhook/0.log" Dec 06 09:41:56 crc kubenswrapper[4945]: I1206 09:41:56.758704 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-546584f4f9-c5crz_11477f2b-5358-45d2-921c-d2ced41fe1f0/prometheus-operator-admission-webhook/0.log" Dec 06 09:41:56 crc kubenswrapper[4945]: I1206 09:41:56.902512 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-twxbc_ccaddde6-c87e-4aba-b71f-4974135fa01a/operator/0.log" Dec 06 09:41:56 crc kubenswrapper[4945]: I1206 09:41:56.983015 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-mg69q_1e66cba6-f587-4390-9f2c-69003f179cae/perses-operator/0.log" Dec 06 09:42:32 crc kubenswrapper[4945]: I1206 09:42:32.553273 4945 scope.go:117] "RemoveContainer" containerID="0c4dafdb4e99fef8763676a8f5b0b180a2ecccc8a12750c9da3f31686cf532cf" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.337730 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qnkzc"] Dec 06 09:43:22 crc kubenswrapper[4945]: E1206 09:43:22.338792 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd9f26a-139e-476b-a850-258ec8da775b" containerName="extract-content" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.338805 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd9f26a-139e-476b-a850-258ec8da775b" containerName="extract-content" Dec 06 09:43:22 crc kubenswrapper[4945]: E1206 09:43:22.338813 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd9f26a-139e-476b-a850-258ec8da775b" containerName="extract-utilities" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.338819 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd9f26a-139e-476b-a850-258ec8da775b" containerName="extract-utilities" Dec 06 09:43:22 crc kubenswrapper[4945]: E1206 09:43:22.338846 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerName="registry-server" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.338853 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerName="registry-server" Dec 06 09:43:22 crc kubenswrapper[4945]: E1206 09:43:22.338869 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerName="extract-content" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.338874 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerName="extract-content" Dec 06 09:43:22 crc kubenswrapper[4945]: E1206 09:43:22.338896 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd9f26a-139e-476b-a850-258ec8da775b" containerName="registry-server" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.338902 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd9f26a-139e-476b-a850-258ec8da775b" containerName="registry-server" Dec 06 09:43:22 crc kubenswrapper[4945]: E1206 09:43:22.338917 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerName="extract-utilities" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.338922 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerName="extract-utilities" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.339100 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4c1fda-a823-4eb3-9ce9-ee374ec2aefa" containerName="registry-server" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.339131 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddd9f26a-139e-476b-a850-258ec8da775b" containerName="registry-server" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.340652 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.349200 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qnkzc"] Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.420835 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-utilities\") pod \"certified-operators-qnkzc\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.420893 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvdpb\" (UniqueName: \"kubernetes.io/projected/e2b05215-d486-40c0-92a8-352f92638fdf-kube-api-access-zvdpb\") pod \"certified-operators-qnkzc\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.420935 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-catalog-content\") pod \"certified-operators-qnkzc\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.523618 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-utilities\") pod \"certified-operators-qnkzc\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.523683 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvdpb\" (UniqueName: \"kubernetes.io/projected/e2b05215-d486-40c0-92a8-352f92638fdf-kube-api-access-zvdpb\") pod \"certified-operators-qnkzc\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.524148 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-catalog-content\") pod \"certified-operators-qnkzc\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.524198 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-catalog-content\") pod \"certified-operators-qnkzc\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.524151 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-utilities\") pod \"certified-operators-qnkzc\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.547919 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvdpb\" (UniqueName: \"kubernetes.io/projected/e2b05215-d486-40c0-92a8-352f92638fdf-kube-api-access-zvdpb\") pod \"certified-operators-qnkzc\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:22 crc kubenswrapper[4945]: I1206 09:43:22.676953 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:23 crc kubenswrapper[4945]: I1206 09:43:23.298547 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qnkzc"] Dec 06 09:43:24 crc kubenswrapper[4945]: I1206 09:43:24.102192 4945 generic.go:334] "Generic (PLEG): container finished" podID="e2b05215-d486-40c0-92a8-352f92638fdf" containerID="66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f" exitCode=0 Dec 06 09:43:24 crc kubenswrapper[4945]: I1206 09:43:24.102472 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qnkzc" event={"ID":"e2b05215-d486-40c0-92a8-352f92638fdf","Type":"ContainerDied","Data":"66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f"} Dec 06 09:43:24 crc kubenswrapper[4945]: I1206 09:43:24.102496 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qnkzc" event={"ID":"e2b05215-d486-40c0-92a8-352f92638fdf","Type":"ContainerStarted","Data":"5c2a5cdb2458e63b0e9a23013dda13a5570105119c892f9328b06a9abe1e9249"} Dec 06 09:43:24 crc kubenswrapper[4945]: I1206 09:43:24.106327 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 09:43:25 crc kubenswrapper[4945]: I1206 09:43:25.115858 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qnkzc" event={"ID":"e2b05215-d486-40c0-92a8-352f92638fdf","Type":"ContainerStarted","Data":"9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f"} Dec 06 09:43:26 crc kubenswrapper[4945]: I1206 09:43:26.129450 4945 generic.go:334] "Generic (PLEG): container finished" podID="e2b05215-d486-40c0-92a8-352f92638fdf" containerID="9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f" exitCode=0 Dec 06 09:43:26 crc kubenswrapper[4945]: I1206 09:43:26.129669 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qnkzc" event={"ID":"e2b05215-d486-40c0-92a8-352f92638fdf","Type":"ContainerDied","Data":"9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f"} Dec 06 09:43:28 crc kubenswrapper[4945]: I1206 09:43:28.153793 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qnkzc" event={"ID":"e2b05215-d486-40c0-92a8-352f92638fdf","Type":"ContainerStarted","Data":"8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509"} Dec 06 09:43:28 crc kubenswrapper[4945]: I1206 09:43:28.191448 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qnkzc" podStartSLOduration=3.694668134 podStartE2EDuration="6.191427788s" podCreationTimestamp="2025-12-06 09:43:22 +0000 UTC" firstStartedPulling="2025-12-06 09:43:24.104328446 +0000 UTC m=+10217.559189500" lastFinishedPulling="2025-12-06 09:43:26.60108811 +0000 UTC m=+10220.055949154" observedRunningTime="2025-12-06 09:43:28.179045757 +0000 UTC m=+10221.633906801" watchObservedRunningTime="2025-12-06 09:43:28.191427788 +0000 UTC m=+10221.646288832" Dec 06 09:43:32 crc kubenswrapper[4945]: I1206 09:43:32.677360 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:32 crc kubenswrapper[4945]: I1206 09:43:32.678200 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:32 crc kubenswrapper[4945]: I1206 09:43:32.735483 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:33 crc kubenswrapper[4945]: I1206 09:43:33.271574 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:33 crc kubenswrapper[4945]: I1206 09:43:33.323299 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qnkzc"] Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.228103 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qnkzc" podUID="e2b05215-d486-40c0-92a8-352f92638fdf" containerName="registry-server" containerID="cri-o://8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509" gracePeriod=2 Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.743880 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.822043 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvdpb\" (UniqueName: \"kubernetes.io/projected/e2b05215-d486-40c0-92a8-352f92638fdf-kube-api-access-zvdpb\") pod \"e2b05215-d486-40c0-92a8-352f92638fdf\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.822169 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-utilities\") pod \"e2b05215-d486-40c0-92a8-352f92638fdf\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.823266 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-utilities" (OuterVolumeSpecName: "utilities") pod "e2b05215-d486-40c0-92a8-352f92638fdf" (UID: "e2b05215-d486-40c0-92a8-352f92638fdf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.823481 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-catalog-content\") pod \"e2b05215-d486-40c0-92a8-352f92638fdf\" (UID: \"e2b05215-d486-40c0-92a8-352f92638fdf\") " Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.827534 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.828397 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2b05215-d486-40c0-92a8-352f92638fdf-kube-api-access-zvdpb" (OuterVolumeSpecName: "kube-api-access-zvdpb") pod "e2b05215-d486-40c0-92a8-352f92638fdf" (UID: "e2b05215-d486-40c0-92a8-352f92638fdf"). InnerVolumeSpecName "kube-api-access-zvdpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.927188 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2b05215-d486-40c0-92a8-352f92638fdf" (UID: "e2b05215-d486-40c0-92a8-352f92638fdf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.929431 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2b05215-d486-40c0-92a8-352f92638fdf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:43:35 crc kubenswrapper[4945]: I1206 09:43:35.929471 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvdpb\" (UniqueName: \"kubernetes.io/projected/e2b05215-d486-40c0-92a8-352f92638fdf-kube-api-access-zvdpb\") on node \"crc\" DevicePath \"\"" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.241222 4945 generic.go:334] "Generic (PLEG): container finished" podID="e2b05215-d486-40c0-92a8-352f92638fdf" containerID="8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509" exitCode=0 Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.241290 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qnkzc" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.241272 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qnkzc" event={"ID":"e2b05215-d486-40c0-92a8-352f92638fdf","Type":"ContainerDied","Data":"8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509"} Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.241370 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qnkzc" event={"ID":"e2b05215-d486-40c0-92a8-352f92638fdf","Type":"ContainerDied","Data":"5c2a5cdb2458e63b0e9a23013dda13a5570105119c892f9328b06a9abe1e9249"} Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.241413 4945 scope.go:117] "RemoveContainer" containerID="8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.260537 4945 scope.go:117] "RemoveContainer" containerID="9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.275781 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qnkzc"] Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.286112 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qnkzc"] Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.560335 4945 scope.go:117] "RemoveContainer" containerID="66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.610607 4945 scope.go:117] "RemoveContainer" containerID="8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509" Dec 06 09:43:36 crc kubenswrapper[4945]: E1206 09:43:36.611075 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509\": container with ID starting with 8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509 not found: ID does not exist" containerID="8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.611108 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509"} err="failed to get container status \"8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509\": rpc error: code = NotFound desc = could not find container \"8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509\": container with ID starting with 8182ea8bf41b0fbb8d6891181938f286de2a6391efe9f3b0300d82672a7ed509 not found: ID does not exist" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.611128 4945 scope.go:117] "RemoveContainer" containerID="9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f" Dec 06 09:43:36 crc kubenswrapper[4945]: E1206 09:43:36.611691 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f\": container with ID starting with 9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f not found: ID does not exist" containerID="9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.611736 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f"} err="failed to get container status \"9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f\": rpc error: code = NotFound desc = could not find container \"9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f\": container with ID starting with 9e57423afd73f063e076a391cd00f0a69904dcc53acd92bf7b94e3fc8be74f7f not found: ID does not exist" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.611763 4945 scope.go:117] "RemoveContainer" containerID="66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f" Dec 06 09:43:36 crc kubenswrapper[4945]: E1206 09:43:36.612301 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f\": container with ID starting with 66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f not found: ID does not exist" containerID="66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.612331 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f"} err="failed to get container status \"66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f\": rpc error: code = NotFound desc = could not find container \"66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f\": container with ID starting with 66f473da9560046bd13214c14c5d6669c69f617802e6d9d69a1529eadb615e7f not found: ID does not exist" Dec 06 09:43:36 crc kubenswrapper[4945]: I1206 09:43:36.973685 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2b05215-d486-40c0-92a8-352f92638fdf" path="/var/lib/kubelet/pods/e2b05215-d486-40c0-92a8-352f92638fdf/volumes" Dec 06 09:43:59 crc kubenswrapper[4945]: E1206 09:43:59.058103 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34ce5b75_9ea5_4856_9975_d0ffd4ca3cf1.slice/crio-c531f85676251f5ba45ec49ca3a92bbbb93c4c8fe11e8535703806fed9954f4a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34ce5b75_9ea5_4856_9975_d0ffd4ca3cf1.slice/crio-conmon-c531f85676251f5ba45ec49ca3a92bbbb93c4c8fe11e8535703806fed9954f4a.scope\": RecentStats: unable to find data in memory cache]" Dec 06 09:43:59 crc kubenswrapper[4945]: I1206 09:43:59.518162 4945 generic.go:334] "Generic (PLEG): container finished" podID="34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" containerID="c531f85676251f5ba45ec49ca3a92bbbb93c4c8fe11e8535703806fed9954f4a" exitCode=0 Dec 06 09:43:59 crc kubenswrapper[4945]: I1206 09:43:59.518240 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rz6dz/must-gather-5r6s7" event={"ID":"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1","Type":"ContainerDied","Data":"c531f85676251f5ba45ec49ca3a92bbbb93c4c8fe11e8535703806fed9954f4a"} Dec 06 09:43:59 crc kubenswrapper[4945]: I1206 09:43:59.519383 4945 scope.go:117] "RemoveContainer" containerID="c531f85676251f5ba45ec49ca3a92bbbb93c4c8fe11e8535703806fed9954f4a" Dec 06 09:44:00 crc kubenswrapper[4945]: I1206 09:44:00.042174 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rz6dz_must-gather-5r6s7_34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1/gather/0.log" Dec 06 09:44:08 crc kubenswrapper[4945]: I1206 09:44:08.795475 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:44:08 crc kubenswrapper[4945]: I1206 09:44:08.796046 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:44:09 crc kubenswrapper[4945]: I1206 09:44:09.320731 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rz6dz/must-gather-5r6s7"] Dec 06 09:44:09 crc kubenswrapper[4945]: I1206 09:44:09.321268 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-rz6dz/must-gather-5r6s7" podUID="34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" containerName="copy" containerID="cri-o://ff36ec5138ea04fe9a79d48e0fa9d78cc51a47f3146157fba2a8023378ebf6ef" gracePeriod=2 Dec 06 09:44:09 crc kubenswrapper[4945]: I1206 09:44:09.338235 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rz6dz/must-gather-5r6s7"] Dec 06 09:44:09 crc kubenswrapper[4945]: I1206 09:44:09.642485 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rz6dz_must-gather-5r6s7_34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1/copy/0.log" Dec 06 09:44:09 crc kubenswrapper[4945]: I1206 09:44:09.643449 4945 generic.go:334] "Generic (PLEG): container finished" podID="34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" containerID="ff36ec5138ea04fe9a79d48e0fa9d78cc51a47f3146157fba2a8023378ebf6ef" exitCode=143 Dec 06 09:44:09 crc kubenswrapper[4945]: I1206 09:44:09.895483 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rz6dz_must-gather-5r6s7_34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1/copy/0.log" Dec 06 09:44:09 crc kubenswrapper[4945]: I1206 09:44:09.896137 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/must-gather-5r6s7" Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.027873 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-must-gather-output\") pod \"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1\" (UID: \"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1\") " Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.028073 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctg7s\" (UniqueName: \"kubernetes.io/projected/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-kube-api-access-ctg7s\") pod \"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1\" (UID: \"34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1\") " Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.034773 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-kube-api-access-ctg7s" (OuterVolumeSpecName: "kube-api-access-ctg7s") pod "34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" (UID: "34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1"). InnerVolumeSpecName "kube-api-access-ctg7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.130801 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctg7s\" (UniqueName: \"kubernetes.io/projected/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-kube-api-access-ctg7s\") on node \"crc\" DevicePath \"\"" Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.202582 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" (UID: "34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.232388 4945 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.656534 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rz6dz_must-gather-5r6s7_34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1/copy/0.log" Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.657056 4945 scope.go:117] "RemoveContainer" containerID="ff36ec5138ea04fe9a79d48e0fa9d78cc51a47f3146157fba2a8023378ebf6ef" Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.657111 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rz6dz/must-gather-5r6s7" Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.686120 4945 scope.go:117] "RemoveContainer" containerID="c531f85676251f5ba45ec49ca3a92bbbb93c4c8fe11e8535703806fed9954f4a" Dec 06 09:44:10 crc kubenswrapper[4945]: I1206 09:44:10.964801 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" path="/var/lib/kubelet/pods/34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1/volumes" Dec 06 09:44:38 crc kubenswrapper[4945]: I1206 09:44:38.795932 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:44:38 crc kubenswrapper[4945]: I1206 09:44:38.796552 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.164482 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn"] Dec 06 09:45:00 crc kubenswrapper[4945]: E1206 09:45:00.165511 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2b05215-d486-40c0-92a8-352f92638fdf" containerName="extract-content" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.165528 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2b05215-d486-40c0-92a8-352f92638fdf" containerName="extract-content" Dec 06 09:45:00 crc kubenswrapper[4945]: E1206 09:45:00.165544 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2b05215-d486-40c0-92a8-352f92638fdf" containerName="registry-server" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.165550 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2b05215-d486-40c0-92a8-352f92638fdf" containerName="registry-server" Dec 06 09:45:00 crc kubenswrapper[4945]: E1206 09:45:00.165561 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" containerName="gather" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.165568 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" containerName="gather" Dec 06 09:45:00 crc kubenswrapper[4945]: E1206 09:45:00.165582 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2b05215-d486-40c0-92a8-352f92638fdf" containerName="extract-utilities" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.165588 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2b05215-d486-40c0-92a8-352f92638fdf" containerName="extract-utilities" Dec 06 09:45:00 crc kubenswrapper[4945]: E1206 09:45:00.165621 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" containerName="copy" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.165627 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" containerName="copy" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.165797 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" containerName="gather" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.165813 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2b05215-d486-40c0-92a8-352f92638fdf" containerName="registry-server" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.165828 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ce5b75-9ea5-4856-9975-d0ffd4ca3cf1" containerName="copy" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.166605 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.168553 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.168685 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.202368 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn"] Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.260652 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5jtg\" (UniqueName: \"kubernetes.io/projected/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-kube-api-access-z5jtg\") pod \"collect-profiles-29416905-cwmhn\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.260698 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-secret-volume\") pod \"collect-profiles-29416905-cwmhn\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.260740 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-config-volume\") pod \"collect-profiles-29416905-cwmhn\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.362617 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-secret-volume\") pod \"collect-profiles-29416905-cwmhn\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.362719 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-config-volume\") pod \"collect-profiles-29416905-cwmhn\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.362920 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5jtg\" (UniqueName: \"kubernetes.io/projected/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-kube-api-access-z5jtg\") pod \"collect-profiles-29416905-cwmhn\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.364505 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-config-volume\") pod \"collect-profiles-29416905-cwmhn\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.372442 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-secret-volume\") pod \"collect-profiles-29416905-cwmhn\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.382434 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5jtg\" (UniqueName: \"kubernetes.io/projected/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-kube-api-access-z5jtg\") pod \"collect-profiles-29416905-cwmhn\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:00 crc kubenswrapper[4945]: I1206 09:45:00.554781 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:01 crc kubenswrapper[4945]: I1206 09:45:01.005258 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn"] Dec 06 09:45:01 crc kubenswrapper[4945]: I1206 09:45:01.195576 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" event={"ID":"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a","Type":"ContainerStarted","Data":"d1dc4e1e0910d1d13d92042b22e67505587460240b0e1060fab6bc64826c58d2"} Dec 06 09:45:02 crc kubenswrapper[4945]: I1206 09:45:02.207653 4945 generic.go:334] "Generic (PLEG): container finished" podID="94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a" containerID="7e8e2a6a6dc297ba1e529b3ac09033c268d61f6f0c1d5c2415e91512b31f4c63" exitCode=0 Dec 06 09:45:02 crc kubenswrapper[4945]: I1206 09:45:02.207702 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" event={"ID":"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a","Type":"ContainerDied","Data":"7e8e2a6a6dc297ba1e529b3ac09033c268d61f6f0c1d5c2415e91512b31f4c63"} Dec 06 09:45:03 crc kubenswrapper[4945]: I1206 09:45:03.581520 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:03 crc kubenswrapper[4945]: I1206 09:45:03.731846 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5jtg\" (UniqueName: \"kubernetes.io/projected/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-kube-api-access-z5jtg\") pod \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " Dec 06 09:45:03 crc kubenswrapper[4945]: I1206 09:45:03.732022 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-config-volume\") pod \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " Dec 06 09:45:03 crc kubenswrapper[4945]: I1206 09:45:03.732138 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-secret-volume\") pod \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\" (UID: \"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a\") " Dec 06 09:45:03 crc kubenswrapper[4945]: I1206 09:45:03.733248 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-config-volume" (OuterVolumeSpecName: "config-volume") pod "94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a" (UID: "94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 09:45:03 crc kubenswrapper[4945]: I1206 09:45:03.736662 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a" (UID: "94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 09:45:03 crc kubenswrapper[4945]: I1206 09:45:03.737402 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-kube-api-access-z5jtg" (OuterVolumeSpecName: "kube-api-access-z5jtg") pod "94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a" (UID: "94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a"). InnerVolumeSpecName "kube-api-access-z5jtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:45:03 crc kubenswrapper[4945]: I1206 09:45:03.834317 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5jtg\" (UniqueName: \"kubernetes.io/projected/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-kube-api-access-z5jtg\") on node \"crc\" DevicePath \"\"" Dec 06 09:45:03 crc kubenswrapper[4945]: I1206 09:45:03.834359 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 09:45:03 crc kubenswrapper[4945]: I1206 09:45:03.834369 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 09:45:04 crc kubenswrapper[4945]: I1206 09:45:04.238206 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" event={"ID":"94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a","Type":"ContainerDied","Data":"d1dc4e1e0910d1d13d92042b22e67505587460240b0e1060fab6bc64826c58d2"} Dec 06 09:45:04 crc kubenswrapper[4945]: I1206 09:45:04.238327 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1dc4e1e0910d1d13d92042b22e67505587460240b0e1060fab6bc64826c58d2" Dec 06 09:45:04 crc kubenswrapper[4945]: I1206 09:45:04.238395 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416905-cwmhn" Dec 06 09:45:04 crc kubenswrapper[4945]: I1206 09:45:04.673952 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj"] Dec 06 09:45:04 crc kubenswrapper[4945]: I1206 09:45:04.683237 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416860-t29bj"] Dec 06 09:45:04 crc kubenswrapper[4945]: I1206 09:45:04.967189 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46722c9b-93e7-4d89-9de7-190570bd94a5" path="/var/lib/kubelet/pods/46722c9b-93e7-4d89-9de7-190570bd94a5/volumes" Dec 06 09:45:08 crc kubenswrapper[4945]: I1206 09:45:08.795641 4945 patch_prober.go:28] interesting pod/machine-config-daemon-57tvp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 09:45:08 crc kubenswrapper[4945]: I1206 09:45:08.796246 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 09:45:08 crc kubenswrapper[4945]: I1206 09:45:08.796304 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" Dec 06 09:45:08 crc kubenswrapper[4945]: I1206 09:45:08.796970 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931"} pod="openshift-machine-config-operator/machine-config-daemon-57tvp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 09:45:08 crc kubenswrapper[4945]: I1206 09:45:08.797018 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerName="machine-config-daemon" containerID="cri-o://50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" gracePeriod=600 Dec 06 09:45:08 crc kubenswrapper[4945]: E1206 09:45:08.924873 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:45:09 crc kubenswrapper[4945]: I1206 09:45:09.288770 4945 generic.go:334] "Generic (PLEG): container finished" podID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" exitCode=0 Dec 06 09:45:09 crc kubenswrapper[4945]: I1206 09:45:09.288818 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" event={"ID":"f56279eb-fe4e-4bd2-bccd-d39958e923ea","Type":"ContainerDied","Data":"50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931"} Dec 06 09:45:09 crc kubenswrapper[4945]: I1206 09:45:09.288853 4945 scope.go:117] "RemoveContainer" containerID="64748f113951c547ded8dabe4d2c5762e6623dd8ea101b8b8c6080bfda5b945a" Dec 06 09:45:09 crc kubenswrapper[4945]: I1206 09:45:09.289542 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:45:09 crc kubenswrapper[4945]: E1206 09:45:09.290162 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:45:21 crc kubenswrapper[4945]: I1206 09:45:21.960235 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:45:21 crc kubenswrapper[4945]: E1206 09:45:21.962249 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:45:32 crc kubenswrapper[4945]: I1206 09:45:32.745849 4945 scope.go:117] "RemoveContainer" containerID="32b8a3775cdb1a9722c999fa6ec938af3fbf2b6d1c5be738e83375f9a21d37c7" Dec 06 09:45:34 crc kubenswrapper[4945]: I1206 09:45:34.953585 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:45:34 crc kubenswrapper[4945]: E1206 09:45:34.954668 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:45:47 crc kubenswrapper[4945]: I1206 09:45:47.953964 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:45:47 crc kubenswrapper[4945]: E1206 09:45:47.954740 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:46:00 crc kubenswrapper[4945]: I1206 09:46:00.953612 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:46:00 crc kubenswrapper[4945]: E1206 09:46:00.954351 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:46:15 crc kubenswrapper[4945]: I1206 09:46:15.952982 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:46:15 crc kubenswrapper[4945]: E1206 09:46:15.953765 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:46:28 crc kubenswrapper[4945]: I1206 09:46:28.954221 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:46:28 crc kubenswrapper[4945]: E1206 09:46:28.955131 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:46:43 crc kubenswrapper[4945]: I1206 09:46:43.957295 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:46:43 crc kubenswrapper[4945]: E1206 09:46:43.958107 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:46:54 crc kubenswrapper[4945]: I1206 09:46:54.954184 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:46:54 crc kubenswrapper[4945]: E1206 09:46:54.956203 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:47:05 crc kubenswrapper[4945]: I1206 09:47:05.953995 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:47:05 crc kubenswrapper[4945]: E1206 09:47:05.954752 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:47:16 crc kubenswrapper[4945]: I1206 09:47:16.961857 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:47:16 crc kubenswrapper[4945]: E1206 09:47:16.962762 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.602382 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8xmmb"] Dec 06 09:47:26 crc kubenswrapper[4945]: E1206 09:47:26.603559 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a" containerName="collect-profiles" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.603577 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a" containerName="collect-profiles" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.603980 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="94ad299d-dc0e-43b4-ba35-cb9efa3a0e6a" containerName="collect-profiles" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.606264 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.617709 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8xmmb"] Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.689445 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsjbt\" (UniqueName: \"kubernetes.io/projected/ec417d82-37b5-49de-8df8-b9c1a360f1cb-kube-api-access-zsjbt\") pod \"community-operators-8xmmb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.689593 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-utilities\") pod \"community-operators-8xmmb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.689611 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-catalog-content\") pod \"community-operators-8xmmb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.791496 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-utilities\") pod \"community-operators-8xmmb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.791547 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-catalog-content\") pod \"community-operators-8xmmb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.791634 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsjbt\" (UniqueName: \"kubernetes.io/projected/ec417d82-37b5-49de-8df8-b9c1a360f1cb-kube-api-access-zsjbt\") pod \"community-operators-8xmmb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.792149 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-utilities\") pod \"community-operators-8xmmb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.792205 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-catalog-content\") pod \"community-operators-8xmmb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.823294 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsjbt\" (UniqueName: \"kubernetes.io/projected/ec417d82-37b5-49de-8df8-b9c1a360f1cb-kube-api-access-zsjbt\") pod \"community-operators-8xmmb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:26 crc kubenswrapper[4945]: I1206 09:47:26.938985 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:27 crc kubenswrapper[4945]: I1206 09:47:27.537508 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8xmmb"] Dec 06 09:47:27 crc kubenswrapper[4945]: I1206 09:47:27.836128 4945 generic.go:334] "Generic (PLEG): container finished" podID="ec417d82-37b5-49de-8df8-b9c1a360f1cb" containerID="aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877" exitCode=0 Dec 06 09:47:27 crc kubenswrapper[4945]: I1206 09:47:27.836202 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xmmb" event={"ID":"ec417d82-37b5-49de-8df8-b9c1a360f1cb","Type":"ContainerDied","Data":"aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877"} Dec 06 09:47:27 crc kubenswrapper[4945]: I1206 09:47:27.836486 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xmmb" event={"ID":"ec417d82-37b5-49de-8df8-b9c1a360f1cb","Type":"ContainerStarted","Data":"b7e7efa000f9f078adbb12020cb5540390ad22c70a0b35e300fa324d0640cf46"} Dec 06 09:47:27 crc kubenswrapper[4945]: I1206 09:47:27.954699 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:47:27 crc kubenswrapper[4945]: E1206 09:47:27.955217 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:47:28 crc kubenswrapper[4945]: I1206 09:47:28.848819 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xmmb" event={"ID":"ec417d82-37b5-49de-8df8-b9c1a360f1cb","Type":"ContainerStarted","Data":"a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01"} Dec 06 09:47:29 crc kubenswrapper[4945]: I1206 09:47:29.865526 4945 generic.go:334] "Generic (PLEG): container finished" podID="ec417d82-37b5-49de-8df8-b9c1a360f1cb" containerID="a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01" exitCode=0 Dec 06 09:47:29 crc kubenswrapper[4945]: I1206 09:47:29.865580 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xmmb" event={"ID":"ec417d82-37b5-49de-8df8-b9c1a360f1cb","Type":"ContainerDied","Data":"a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01"} Dec 06 09:47:30 crc kubenswrapper[4945]: I1206 09:47:30.878584 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xmmb" event={"ID":"ec417d82-37b5-49de-8df8-b9c1a360f1cb","Type":"ContainerStarted","Data":"90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f"} Dec 06 09:47:30 crc kubenswrapper[4945]: I1206 09:47:30.903111 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8xmmb" podStartSLOduration=2.445261935 podStartE2EDuration="4.903092302s" podCreationTimestamp="2025-12-06 09:47:26 +0000 UTC" firstStartedPulling="2025-12-06 09:47:27.838035933 +0000 UTC m=+10461.292896977" lastFinishedPulling="2025-12-06 09:47:30.2958663 +0000 UTC m=+10463.750727344" observedRunningTime="2025-12-06 09:47:30.897512222 +0000 UTC m=+10464.352373276" watchObservedRunningTime="2025-12-06 09:47:30.903092302 +0000 UTC m=+10464.357953346" Dec 06 09:47:36 crc kubenswrapper[4945]: I1206 09:47:36.939487 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:36 crc kubenswrapper[4945]: I1206 09:47:36.940155 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:36 crc kubenswrapper[4945]: I1206 09:47:36.989364 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:37 crc kubenswrapper[4945]: I1206 09:47:37.995486 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:38 crc kubenswrapper[4945]: I1206 09:47:38.046081 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8xmmb"] Dec 06 09:47:39 crc kubenswrapper[4945]: I1206 09:47:39.970356 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8xmmb" podUID="ec417d82-37b5-49de-8df8-b9c1a360f1cb" containerName="registry-server" containerID="cri-o://90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f" gracePeriod=2 Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.520036 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.617380 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-utilities\") pod \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.617519 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-catalog-content\") pod \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.617545 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsjbt\" (UniqueName: \"kubernetes.io/projected/ec417d82-37b5-49de-8df8-b9c1a360f1cb-kube-api-access-zsjbt\") pod \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\" (UID: \"ec417d82-37b5-49de-8df8-b9c1a360f1cb\") " Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.618662 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-utilities" (OuterVolumeSpecName: "utilities") pod "ec417d82-37b5-49de-8df8-b9c1a360f1cb" (UID: "ec417d82-37b5-49de-8df8-b9c1a360f1cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.623443 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec417d82-37b5-49de-8df8-b9c1a360f1cb-kube-api-access-zsjbt" (OuterVolumeSpecName: "kube-api-access-zsjbt") pod "ec417d82-37b5-49de-8df8-b9c1a360f1cb" (UID: "ec417d82-37b5-49de-8df8-b9c1a360f1cb"). InnerVolumeSpecName "kube-api-access-zsjbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.684186 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec417d82-37b5-49de-8df8-b9c1a360f1cb" (UID: "ec417d82-37b5-49de-8df8-b9c1a360f1cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.719872 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.719912 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec417d82-37b5-49de-8df8-b9c1a360f1cb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.719926 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsjbt\" (UniqueName: \"kubernetes.io/projected/ec417d82-37b5-49de-8df8-b9c1a360f1cb-kube-api-access-zsjbt\") on node \"crc\" DevicePath \"\"" Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.980017 4945 generic.go:334] "Generic (PLEG): container finished" podID="ec417d82-37b5-49de-8df8-b9c1a360f1cb" containerID="90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f" exitCode=0 Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.980063 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xmmb" Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.980065 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xmmb" event={"ID":"ec417d82-37b5-49de-8df8-b9c1a360f1cb","Type":"ContainerDied","Data":"90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f"} Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.980174 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xmmb" event={"ID":"ec417d82-37b5-49de-8df8-b9c1a360f1cb","Type":"ContainerDied","Data":"b7e7efa000f9f078adbb12020cb5540390ad22c70a0b35e300fa324d0640cf46"} Dec 06 09:47:40 crc kubenswrapper[4945]: I1206 09:47:40.980198 4945 scope.go:117] "RemoveContainer" containerID="90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f" Dec 06 09:47:41 crc kubenswrapper[4945]: I1206 09:47:41.016321 4945 scope.go:117] "RemoveContainer" containerID="a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01" Dec 06 09:47:41 crc kubenswrapper[4945]: I1206 09:47:41.028152 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8xmmb"] Dec 06 09:47:41 crc kubenswrapper[4945]: I1206 09:47:41.037921 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8xmmb"] Dec 06 09:47:41 crc kubenswrapper[4945]: I1206 09:47:41.068729 4945 scope.go:117] "RemoveContainer" containerID="aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877" Dec 06 09:47:41 crc kubenswrapper[4945]: I1206 09:47:41.090691 4945 scope.go:117] "RemoveContainer" containerID="90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f" Dec 06 09:47:41 crc kubenswrapper[4945]: E1206 09:47:41.091227 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f\": container with ID starting with 90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f not found: ID does not exist" containerID="90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f" Dec 06 09:47:41 crc kubenswrapper[4945]: I1206 09:47:41.091272 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f"} err="failed to get container status \"90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f\": rpc error: code = NotFound desc = could not find container \"90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f\": container with ID starting with 90340b6e551850212c1d04ecd4d076fd09cba979190e5100b75f4c81cc19bb1f not found: ID does not exist" Dec 06 09:47:41 crc kubenswrapper[4945]: I1206 09:47:41.091344 4945 scope.go:117] "RemoveContainer" containerID="a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01" Dec 06 09:47:41 crc kubenswrapper[4945]: E1206 09:47:41.091794 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01\": container with ID starting with a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01 not found: ID does not exist" containerID="a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01" Dec 06 09:47:41 crc kubenswrapper[4945]: I1206 09:47:41.091830 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01"} err="failed to get container status \"a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01\": rpc error: code = NotFound desc = could not find container \"a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01\": container with ID starting with a90a1525a8a7b7b12349a812e1944e7152b329be49d492ad329adfdf9e93bb01 not found: ID does not exist" Dec 06 09:47:41 crc kubenswrapper[4945]: I1206 09:47:41.091851 4945 scope.go:117] "RemoveContainer" containerID="aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877" Dec 06 09:47:41 crc kubenswrapper[4945]: E1206 09:47:41.092349 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877\": container with ID starting with aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877 not found: ID does not exist" containerID="aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877" Dec 06 09:47:41 crc kubenswrapper[4945]: I1206 09:47:41.092393 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877"} err="failed to get container status \"aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877\": rpc error: code = NotFound desc = could not find container \"aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877\": container with ID starting with aeed6a917ef1ab8998633d3d3f10e70b520dac9b59d443d3c1d55fbf52e33877 not found: ID does not exist" Dec 06 09:47:42 crc kubenswrapper[4945]: I1206 09:47:42.955478 4945 scope.go:117] "RemoveContainer" containerID="50ea8b12a030f0ea94ecf021a3ff6fdcf6d77eaded15f109a7364eb15a78e931" Dec 06 09:47:42 crc kubenswrapper[4945]: E1206 09:47:42.956317 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-57tvp_openshift-machine-config-operator(f56279eb-fe4e-4bd2-bccd-d39958e923ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-57tvp" podUID="f56279eb-fe4e-4bd2-bccd-d39958e923ea" Dec 06 09:47:42 crc kubenswrapper[4945]: I1206 09:47:42.965654 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec417d82-37b5-49de-8df8-b9c1a360f1cb" path="/var/lib/kubelet/pods/ec417d82-37b5-49de-8df8-b9c1a360f1cb/volumes"